[ 456.058321] env[63418]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63418) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 456.058661] env[63418]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63418) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 456.058705] env[63418]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63418) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 456.059075] env[63418]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 456.153898] env[63418]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63418) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 456.163772] env[63418]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63418) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 456.768578] env[63418]: INFO nova.virt.driver [None req-6644e034-fa75-4f62-a3cd-7e4fc67cbf44 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 456.840413] env[63418]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 456.840635] env[63418]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 456.840703] env[63418]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63418) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 459.940448] env[63418]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-cc3cce23-2c5a-40f8-9ae2-eeab945b1953 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 459.956830] env[63418]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63418) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 459.957073] env[63418]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-05584a31-25d2-45ab-b0e0-f1c6054b07b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 459.999224] env[63418]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 1d8c3. [ 459.999425] env[63418]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.159s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 459.999909] env[63418]: INFO nova.virt.vmwareapi.driver [None req-6644e034-fa75-4f62-a3cd-7e4fc67cbf44 None None] VMware vCenter version: 7.0.3 [ 460.003404] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144af71d-d563-4ac1-9a2e-6bc6890c44ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.021070] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3454fa8-b2a0-4928-b05d-bcc4f545fb3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.026856] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec6260b-2d16-4eb2-b633-cb3d46cb8c46 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.033177] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f0d6ce-853c-4f3e-9509-a8b3248dfbc4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.045947] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0313880-fcbb-4264-b384-a59dc5c3bd85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.051779] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1693c43c-94d6-4c68-8e8a-d8655d7b7cd6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.081810] env[63418]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-d9559692-7b54-42eb-b7db-3e34bd6a5d5f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.086856] env[63418]: DEBUG nova.virt.vmwareapi.driver [None req-6644e034-fa75-4f62-a3cd-7e4fc67cbf44 None None] Extension org.openstack.compute already exists. {{(pid=63418) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 460.089604] env[63418]: INFO nova.compute.provider_config [None req-6644e034-fa75-4f62-a3cd-7e4fc67cbf44 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 460.593121] env[63418]: DEBUG nova.context [None req-6644e034-fa75-4f62-a3cd-7e4fc67cbf44 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),5d9288f7-f9aa-44dc-9618-16f0b330eead(cell1) {{(pid=63418) load_cells /opt/stack/nova/nova/context.py:464}} [ 460.595275] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 460.595501] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 460.596177] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 460.596610] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Acquiring lock "5d9288f7-f9aa-44dc-9618-16f0b330eead" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 460.596806] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Lock "5d9288f7-f9aa-44dc-9618-16f0b330eead" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 460.597815] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Lock "5d9288f7-f9aa-44dc-9618-16f0b330eead" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 460.617702] env[63418]: INFO dbcounter [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Registered counter for database nova_cell0 [ 460.625973] env[63418]: INFO dbcounter [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Registered counter for database nova_cell1 [ 460.629115] env[63418]: DEBUG oslo_db.sqlalchemy.engines [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63418) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 460.629468] env[63418]: DEBUG oslo_db.sqlalchemy.engines [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63418) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 460.634275] env[63418]: ERROR nova.db.main.api [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 460.634275] env[63418]: result = function(*args, **kwargs) [ 460.634275] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 460.634275] env[63418]: return func(*args, **kwargs) [ 460.634275] env[63418]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 460.634275] env[63418]: result = fn(*args, **kwargs) [ 460.634275] env[63418]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 460.634275] env[63418]: return f(*args, **kwargs) [ 460.634275] env[63418]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 460.634275] env[63418]: return db.service_get_minimum_version(context, binaries) [ 460.634275] env[63418]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 460.634275] env[63418]: _check_db_access() [ 460.634275] env[63418]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 460.634275] env[63418]: stacktrace = ''.join(traceback.format_stack()) [ 460.634275] env[63418]: [ 460.635110] env[63418]: ERROR nova.db.main.api [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 460.635110] env[63418]: result = function(*args, **kwargs) [ 460.635110] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 460.635110] env[63418]: return func(*args, **kwargs) [ 460.635110] env[63418]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 460.635110] env[63418]: result = fn(*args, **kwargs) [ 460.635110] env[63418]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 460.635110] env[63418]: return f(*args, **kwargs) [ 460.635110] env[63418]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 460.635110] env[63418]: return db.service_get_minimum_version(context, binaries) [ 460.635110] env[63418]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 460.635110] env[63418]: _check_db_access() [ 460.635110] env[63418]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 460.635110] env[63418]: stacktrace = ''.join(traceback.format_stack()) [ 460.635110] env[63418]: [ 460.635630] env[63418]: WARNING nova.objects.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 460.635630] env[63418]: WARNING nova.objects.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Failed to get minimum service version for cell 5d9288f7-f9aa-44dc-9618-16f0b330eead [ 460.636039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Acquiring lock "singleton_lock" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 460.636236] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Acquired lock "singleton_lock" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 460.636437] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Releasing lock "singleton_lock" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 460.636757] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Full set of CONF: {{(pid=63418) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 460.636902] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ******************************************************************************** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 460.637038] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Configuration options gathered from: {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 460.637188] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 460.637374] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 460.637501] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ================================================================================ {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 460.637731] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] allow_resize_to_same_host = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.637962] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] arq_binding_timeout = 300 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.638116] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] backdoor_port = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.638247] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] backdoor_socket = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.638413] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] block_device_allocate_retries = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.638576] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] block_device_allocate_retries_interval = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.638741] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cert = self.pem {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.638905] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.639090] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute_monitors = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.639266] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] config_dir = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.639435] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] config_drive_format = iso9660 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.639573] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.639742] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] config_source = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.639905] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] console_host = devstack {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.640084] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] control_exchange = nova {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.640246] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cpu_allocation_ratio = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.640435] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] daemon = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.640574] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] debug = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.640733] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] default_access_ip_network_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.640896] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] default_availability_zone = nova {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.641063] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] default_ephemeral_format = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.641226] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] default_green_pool_size = 1000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.641457] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.641623] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] default_schedule_zone = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.641779] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] disk_allocation_ratio = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.641938] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] enable_new_services = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.642124] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] enabled_apis = ['osapi_compute'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.642290] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] enabled_ssl_apis = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.642450] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] flat_injected = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.642644] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] force_config_drive = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.642807] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] force_raw_images = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.642978] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] graceful_shutdown_timeout = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.643157] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] heal_instance_info_cache_interval = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.643369] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] host = cpu-1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.643563] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.643739] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.643902] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.644126] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.644290] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_build_timeout = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.644466] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_delete_interval = 300 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.644616] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_format = [instance: %(uuid)s] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.644783] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_name_template = instance-%08x {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.644945] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_usage_audit = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.645125] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_usage_audit_period = month {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.645289] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.645452] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.645619] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] internal_service_availability_zone = internal {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.645783] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] key = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.645943] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] live_migration_retry_count = 30 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.646120] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_color = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.646289] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_config_append = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.646446] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.646602] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_dir = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.646758] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.646886] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_options = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647054] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_rotate_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647221] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_rotate_interval_type = days {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647382] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] log_rotation_type = none {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647510] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647636] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647799] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.647960] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.648098] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.648260] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] long_rpc_timeout = 1800 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.648417] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] max_concurrent_builds = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.648574] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] max_concurrent_live_migrations = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.648729] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] max_concurrent_snapshots = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.648887] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] max_local_block_devices = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.649050] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] max_logfile_count = 30 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.649210] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] max_logfile_size_mb = 200 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.649365] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] maximum_instance_delete_attempts = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.649529] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metadata_listen = 0.0.0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.649693] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metadata_listen_port = 8775 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.649855] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metadata_workers = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.650023] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] migrate_max_retries = -1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.650188] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] mkisofs_cmd = genisoimage {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.650431] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.650525] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] my_ip = 10.180.1.21 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.650724] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.650885] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] network_allocate_retries = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.651074] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.651243] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.651403] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] osapi_compute_listen_port = 8774 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.651566] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] osapi_compute_unique_server_name_scope = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.651736] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] osapi_compute_workers = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.651895] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] password_length = 12 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.652062] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] periodic_enable = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.652223] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] periodic_fuzzy_delay = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.652386] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] pointer_model = usbtablet {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.652569] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] preallocate_images = none {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.652739] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] publish_errors = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.652869] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] pybasedir = /opt/stack/nova {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.653035] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ram_allocation_ratio = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.653199] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] rate_limit_burst = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.653365] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] rate_limit_except_level = CRITICAL {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.653542] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] rate_limit_interval = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.653725] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reboot_timeout = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.653887] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reclaim_instance_interval = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.654051] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] record = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.654224] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reimage_timeout_per_gb = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.654388] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] report_interval = 120 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.654547] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] rescue_timeout = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.654707] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reserved_host_cpus = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.654866] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reserved_host_disk_mb = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655031] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reserved_host_memory_mb = 512 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655193] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] reserved_huge_pages = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655351] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] resize_confirm_window = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655507] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] resize_fs_using_block_device = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655665] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] resume_guests_state_on_host_boot = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655829] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.655988] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] rpc_response_timeout = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.656162] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] run_external_periodic_tasks = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.656329] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] running_deleted_instance_action = reap {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.656499] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.656642] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] running_deleted_instance_timeout = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.656802] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler_instance_sync_interval = 120 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.656969] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_down_time = 720 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.657145] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] servicegroup_driver = db {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.657299] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] shell_completion = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.657457] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] shelved_offload_time = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.657614] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] shelved_poll_interval = 3600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.657781] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] shutdown_timeout = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.657940] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] source_is_ipv6 = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.658109] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ssl_only = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.658350] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.658516] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] sync_power_state_interval = 600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.658677] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] sync_power_state_pool_size = 1000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.658840] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] syslog_log_facility = LOG_USER {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.658997] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] tempdir = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.659169] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] timeout_nbd = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.659333] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] transport_url = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.659492] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] update_resources_interval = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.659649] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_cow_images = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.659809] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_eventlog = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.659966] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_journal = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.660137] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_json = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.660298] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_rootwrap_daemon = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.660455] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_stderr = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.660611] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] use_syslog = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.660766] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vcpu_pin_set = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.660928] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plugging_is_fatal = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.661104] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plugging_timeout = 300 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.661269] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] virt_mkfs = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.661426] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] volume_usage_poll_interval = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.661582] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] watch_log_file = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.661746] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] web = /usr/share/spice-html5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 460.661926] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.662101] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.662267] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.662432] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_concurrency.disable_process_locking = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.662735] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.662920] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.663097] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.663269] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.663437] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.663625] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.663816] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.auth_strategy = keystone {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.663981] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.compute_link_prefix = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.664167] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.664339] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.dhcp_domain = novalocal {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.664509] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.enable_instance_password = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.664671] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.glance_link_prefix = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.664837] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665010] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665180] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.instance_list_per_project_cells = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665338] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.list_records_by_skipping_down_cells = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665497] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.local_metadata_per_cell = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665662] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.max_limit = 1000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665828] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.metadata_cache_expiration = 15 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.665998] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.neutron_default_tenant_id = default {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.666185] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.response_validation = warn {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.666351] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.use_neutron_default_nets = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.666527] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.666693] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.666858] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.667038] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.667213] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_dynamic_targets = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.667373] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_jsonfile_path = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.667547] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.667735] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.backend = dogpile.cache.memcached {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.667899] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.backend_argument = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.668079] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.config_prefix = cache.oslo {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.668249] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.dead_timeout = 60.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.668410] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.debug_cache_backend = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.668568] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.enable_retry_client = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.668726] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.enable_socket_keepalive = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.668891] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.enabled = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.669064] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.enforce_fips_mode = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.669225] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.expiration_time = 600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.669386] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.hashclient_retry_attempts = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.669549] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.669713] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_dead_retry = 300 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.669869] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_password = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.670041] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.670204] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.670364] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_pool_maxsize = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.670522] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.670683] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_sasl_enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.670855] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.671032] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.671195] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.memcache_username = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.671357] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.proxies = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.671519] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_db = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.671676] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_password = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.671841] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.672015] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.672188] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_server = localhost:6379 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.672349] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_socket_timeout = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.672507] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.redis_username = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.672691] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.retry_attempts = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.672860] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.retry_delay = 0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.673030] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.socket_keepalive_count = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.673193] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.socket_keepalive_idle = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.673353] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.socket_keepalive_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.673520] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.tls_allowed_ciphers = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.673702] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.tls_cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.673861] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.tls_certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.674032] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.tls_enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.674196] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cache.tls_keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.674365] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.674537] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.auth_type = password {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.674699] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.674893] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.675078] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.675248] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.675411] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.cross_az_attach = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.675573] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.debug = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.675735] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.endpoint_template = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.675899] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.http_retries = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.676072] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.676232] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.676397] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.os_region_name = RegionOne {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.676558] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.676716] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cinder.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.676884] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.677050] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.cpu_dedicated_set = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.677209] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.cpu_shared_set = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.677372] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.image_type_exclude_list = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.677533] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.677698] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.677858] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.678025] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.678198] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.678361] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.resource_provider_association_refresh = 300 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.678521] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.678683] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.shutdown_retry_interval = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.678861] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.679046] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] conductor.workers = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.679227] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] console.allowed_origins = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.679386] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] console.ssl_ciphers = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.679555] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] console.ssl_minimum_version = default {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.679723] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] consoleauth.enforce_session_timeout = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.679889] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] consoleauth.token_ttl = 600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.680069] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.680229] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.680396] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.680550] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.680708] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.680862] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681033] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681195] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681351] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681509] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681666] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681820] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.681975] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.682154] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.service_type = accelerator {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.682316] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.682473] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.682657] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.682822] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.683007] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.683176] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] cyborg.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.683354] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.backend = sqlalchemy {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.683536] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.connection = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.683719] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.connection_debug = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.683888] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.connection_parameters = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.684064] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.connection_recycle_time = 3600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.684229] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.connection_trace = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.684449] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.db_inc_retry_interval = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.684638] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.db_max_retries = 20 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.684807] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.db_max_retry_interval = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.684971] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.db_retry_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.685148] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.max_overflow = 50 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.685311] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.max_pool_size = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.685472] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.max_retries = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.685641] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.685802] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.mysql_wsrep_sync_wait = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.685957] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.pool_timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.686131] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.retry_interval = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.686291] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.slave_connection = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.686450] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.sqlite_synchronous = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.686607] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] database.use_db_reconnect = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.686777] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.backend = sqlalchemy {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.686941] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.connection = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.687115] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.connection_debug = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.687282] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.connection_parameters = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.687443] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.connection_recycle_time = 3600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.687605] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.connection_trace = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.687766] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.db_inc_retry_interval = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.687925] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.db_max_retries = 20 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.688097] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.db_max_retry_interval = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.688261] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.db_retry_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.688420] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.max_overflow = 50 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.688580] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.max_pool_size = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.688740] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.max_retries = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.688903] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.689070] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.689275] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.pool_timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.689387] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.retry_interval = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.689540] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.slave_connection = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.689699] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] api_database.sqlite_synchronous = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.689871] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] devices.enabled_mdev_types = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.690055] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.690228] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.690389] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ephemeral_storage_encryption.enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.690556] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.690716] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.api_servers = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.690878] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.691046] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.691211] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.691369] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.691527] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.691688] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.debug = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.691852] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.default_trusted_certificate_ids = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.692022] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.enable_certificate_validation = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.692190] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.enable_rbd_download = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.692346] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.692515] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.692698] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.692862] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.693031] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.693200] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.num_retries = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.693368] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.rbd_ceph_conf = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.693546] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.rbd_connect_timeout = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.693728] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.rbd_pool = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.693892] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.rbd_user = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.694057] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.694220] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.694376] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.694543] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.service_type = image {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.694712] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.694892] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.695075] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.695239] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.695418] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.695579] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.verify_glance_signatures = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.695738] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] glance.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.695901] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] guestfs.debug = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.696079] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.696244] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.auth_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.696401] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.696557] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.696719] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.696873] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697036] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697197] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697359] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697514] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697670] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697854] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.697977] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.698147] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.698301] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.698467] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.service_type = shared-file-system {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.698627] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.share_apply_policy_timeout = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.698788] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.698944] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.699112] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.699269] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.699443] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.699601] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] manila.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.699767] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] mks.enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.700127] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.700319] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] image_cache.manager_interval = 2400 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.700490] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] image_cache.precache_concurrency = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.700658] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] image_cache.remove_unused_base_images = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.700827] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.700993] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.701181] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] image_cache.subdirectory_name = _base {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.701357] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.api_max_retries = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.701520] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.api_retry_interval = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.701680] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.701841] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.auth_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.701998] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.702161] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.702318] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.702478] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.conductor_group = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.702663] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.702824] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.702979] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.703151] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.703305] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.703462] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.703618] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.703784] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.peer_list = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.703939] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.704109] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.704273] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.serial_console_state_timeout = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.704429] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.704593] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.service_type = baremetal {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.704750] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.shard = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.704931] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.705108] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.705269] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.705425] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.705600] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.705762] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ironic.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.705941] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.706125] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] key_manager.fixed_key = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.706305] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.706465] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.barbican_api_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.706620] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.barbican_endpoint = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.706788] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.barbican_endpoint_type = public {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.706963] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.barbican_region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.707105] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.707261] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.707426] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.707585] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.707741] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.707899] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.number_of_retries = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.708070] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.retry_delay = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.708234] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.send_service_user_token = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.708394] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.708551] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.708710] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.verify_ssl = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.708865] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican.verify_ssl_path = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.709075] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.709292] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.auth_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.709484] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.709619] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.709787] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.709951] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.710122] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.710284] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.710445] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] barbican_service_user.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.710610] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.approle_role_id = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.710766] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.approle_secret_id = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.710932] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.kv_mountpoint = secret {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.711099] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.kv_path = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.711263] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.kv_version = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.711418] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.namespace = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.711573] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.root_token_id = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.711729] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.ssl_ca_crt_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.711892] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.timeout = 60.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.712064] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.use_ssl = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.712232] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.712403] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.712586] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.auth_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.712756] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.712917] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.713087] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.713248] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.713404] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.713559] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.713737] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.713896] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.714061] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.714217] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.714372] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.714528] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.714685] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.714864] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.service_type = identity {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.715056] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.715221] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.715380] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.715536] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.715713] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.715868] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] keystone.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.716072] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.connection_uri = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.716234] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_mode = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.716397] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.716562] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_models = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.716728] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_power_governor_high = performance {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.716891] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.717059] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_power_management = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.717226] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.717385] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.device_detach_attempts = 8 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.717545] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.device_detach_timeout = 20 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.717707] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.disk_cachemodes = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.717861] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.disk_prefix = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.718061] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.enabled_perf_events = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.718202] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.file_backed_memory = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.718368] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.gid_maps = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.718527] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.hw_disk_discard = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.718687] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.hw_machine_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.718856] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_rbd_ceph_conf = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.719033] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.719208] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.719377] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_rbd_glance_store_name = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.719593] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_rbd_pool = rbd {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.719709] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_type = default {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.719867] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.images_volume_group = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.720039] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.inject_key = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.720206] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.inject_partition = -2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.720367] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.inject_password = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.720527] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.iscsi_iface = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.720688] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.iser_use_multipath = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.720849] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721013] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721179] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_downtime = 500 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721339] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721500] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721657] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_inbound_addr = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721819] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.721977] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.722190] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_scheme = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.722386] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_timeout_action = abort {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.722574] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_tunnelled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.722746] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_uri = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.722910] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.live_migration_with_native_tls = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.723082] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.max_queues = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.723251] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.723484] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.723649] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.nfs_mount_options = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.723933] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.724118] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.724286] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.724446] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.724610] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.724773] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.num_pcie_ports = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.724960] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.725143] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.pmem_namespaces = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.725304] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.quobyte_client_cfg = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.725577] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.725750] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.725912] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.726085] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.726247] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rbd_secret_uuid = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.726402] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rbd_user = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.726564] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.726734] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.726891] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rescue_image_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.727055] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rescue_kernel_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.727216] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rescue_ramdisk_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.727382] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.727541] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.rx_queue_size = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.727707] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.smbfs_mount_options = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.727973] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.728169] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.snapshot_compression = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.728317] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.snapshot_image_format = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.728531] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.728698] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.sparse_logical_volumes = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.728862] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.swtpm_enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.729038] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.swtpm_group = tss {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.729210] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.swtpm_user = tss {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.729375] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.sysinfo_serial = unique {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.729534] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.tb_cache_size = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.729692] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.tx_queue_size = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.729854] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.uid_maps = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.730024] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.use_virtio_for_bridges = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.730196] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.virt_type = kvm {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.730365] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.volume_clear = zero {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.730529] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.volume_clear_size = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.730712] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.volume_use_multipath = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.730847] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_cache_path = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.731019] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.731188] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.731350] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.731513] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.731779] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.731950] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.vzstorage_mount_user = stack {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.732128] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.732301] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.732474] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.auth_type = password {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.732662] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.732826] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.732987] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.733156] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.733314] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.733483] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.default_floating_pool = public {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.733682] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.733860] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.extension_sync_interval = 600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.734036] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.http_retries = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.734202] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.734358] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.734516] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.734711] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.734852] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.735048] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.ovs_bridge = br-int {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.735222] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.physnets = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.735392] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.region_name = RegionOne {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.735552] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.735722] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.service_metadata_proxy = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.735880] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.736054] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.service_type = network {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.736219] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.736376] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.736531] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.736687] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.736862] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.737035] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] neutron.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.737246] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] notifications.bdms_in_notifications = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.737379] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] notifications.default_level = INFO {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.737549] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] notifications.notification_format = unversioned {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.737710] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] notifications.notify_on_state_change = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.737881] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.738064] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] pci.alias = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.738275] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] pci.device_spec = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.738393] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] pci.report_in_placement = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.738566] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.738739] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.auth_type = password {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.738908] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.739079] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.739241] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.739404] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.739562] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.739725] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.739877] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.default_domain_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.740057] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.default_domain_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.740225] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.domain_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.740383] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.domain_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.740536] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.740695] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.740849] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.741061] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.741172] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.741338] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.password = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.741496] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.project_domain_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.741660] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.project_domain_name = Default {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.741858] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.project_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.742059] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.project_name = service {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.742232] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.region_name = RegionOne {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.742395] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.742575] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.742760] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.service_type = placement {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.742924] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.743093] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.743255] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.743412] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.system_scope = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.743590] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.743760] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.trust_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.743918] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.user_domain_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.744162] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.user_domain_name = Default {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.744371] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.user_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.744558] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.username = nova {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.744742] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.744904] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] placement.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.745091] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.cores = 20 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.745258] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.count_usage_from_placement = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.745426] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.745597] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.injected_file_content_bytes = 10240 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.745763] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.injected_file_path_length = 255 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.745928] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.injected_files = 5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.746103] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.instances = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.746267] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.key_pairs = 100 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.746429] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.metadata_items = 128 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.746591] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.ram = 51200 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.746752] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.recheck_quota = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.746915] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.server_group_members = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.747089] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] quota.server_groups = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.747263] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.747426] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.747584] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.image_metadata_prefilter = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.747754] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.747921] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.max_attempts = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.748093] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.max_placement_results = 1000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.748259] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.748419] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.748574] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.748745] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] scheduler.workers = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.748915] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.749093] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.749273] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.749442] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.749607] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.749771] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.749933] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.750129] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.750300] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.host_subset_size = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.750464] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.750622] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.750783] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.750943] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.isolated_hosts = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.751155] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.isolated_images = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.751276] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.751435] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.751625] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.751796] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.pci_in_placement = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.751957] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.752126] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.752289] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.752460] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.752646] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.752816] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.752978] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.track_instance_changes = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.753168] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.753336] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metrics.required = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.753500] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metrics.weight_multiplier = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.753683] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.753862] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] metrics.weight_setting = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.754188] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.754367] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] serial_console.enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.754542] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] serial_console.port_range = 10000:20000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.754712] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.754908] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.755157] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] serial_console.serialproxy_port = 6083 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.755343] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.755519] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.auth_type = password {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.755687] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.755847] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756025] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756182] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756338] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756506] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.send_service_user_token = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756669] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756830] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] service_user.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.756997] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.agent_enabled = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.757172] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.757475] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.757667] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.757837] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.html5proxy_port = 6082 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.757997] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.image_compression = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.758168] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.jpeg_compression = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.758326] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.playback_compression = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.758488] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.require_secure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.758651] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.server_listen = 127.0.0.1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.758816] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.758971] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.streaming_mode = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.759215] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] spice.zlib_compression = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.759408] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] upgrade_levels.baseapi = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.759584] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] upgrade_levels.compute = auto {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.759746] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] upgrade_levels.conductor = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.759905] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] upgrade_levels.scheduler = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.760086] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.760250] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.760408] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.760565] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.760730] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.760887] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.761051] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.761252] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.761368] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vendordata_dynamic_auth.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.761541] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.api_retry_count = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.761699] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.ca_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.761868] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.762041] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.cluster_name = testcl1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.762207] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.connection_pool_size = 10 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.762365] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.console_delay_seconds = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.762547] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.datastore_regex = ^datastore.* {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.762753] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.762927] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.host_password = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.763108] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.host_port = 443 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.763279] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.host_username = administrator@vsphere.local {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.763447] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.insecure = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.763623] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.integration_bridge = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.763820] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.maximum_objects = 100 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.763983] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.pbm_default_policy = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.764161] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.pbm_enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.764319] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.pbm_wsdl_location = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.764487] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.764646] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.serial_port_proxy_uri = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.764804] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.serial_port_service_uri = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.764999] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.task_poll_interval = 0.5 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.765192] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.use_linked_clone = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.765361] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.vnc_keymap = en-us {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.765527] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.vnc_port = 5900 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.765692] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vmware.vnc_port_total = 10000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.765875] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.auth_schemes = ['none'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.766061] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.766359] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.766542] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.766715] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.novncproxy_port = 6080 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.766889] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.server_listen = 127.0.0.1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.767067] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.767230] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.vencrypt_ca_certs = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.767389] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.vencrypt_client_cert = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.767551] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vnc.vencrypt_client_key = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.767724] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.767890] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.disable_deep_image_inspection = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.768064] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.768230] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.768389] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.768597] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.disable_rootwrap = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.768706] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.enable_numa_live_migration = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.768864] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.769031] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.769229] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.769443] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.libvirt_disable_apic = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.769613] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.769779] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.769939] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.770132] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.770277] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.770436] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.770596] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.770754] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.770913] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.771089] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.771277] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.771448] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.client_socket_timeout = 900 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.771613] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.default_pool_size = 1000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.771780] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.keep_alive = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.771944] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.max_header_line = 16384 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.772117] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.772280] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.ssl_ca_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.772438] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.ssl_cert_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.772623] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.ssl_key_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.772791] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.tcp_keepidle = 600 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.772979] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.773214] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] zvm.ca_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.773388] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] zvm.cloud_connector_url = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.773707] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.773892] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] zvm.reachable_timeout = 300 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.774088] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.enforce_new_defaults = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.774466] env[63418]: WARNING oslo_config.cfg [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 460.774651] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.enforce_scope = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.774830] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.policy_default_rule = default {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.775022] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.775198] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.policy_file = policy.yaml {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.775369] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.775530] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.775691] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.775849] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.776017] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.776202] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.776378] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.776555] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.connection_string = messaging:// {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.776726] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.enabled = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.776895] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.es_doc_type = notification {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.777070] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.es_scroll_size = 10000 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.777240] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.es_scroll_time = 2m {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.777404] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.filter_error_trace = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.777571] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.hmac_keys = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.777740] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.sentinel_service_name = mymaster {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.777905] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.socket_timeout = 0.1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.778077] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.trace_requests = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.778241] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler.trace_sqlalchemy = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.778420] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler_jaeger.process_tags = {} {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.778581] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler_jaeger.service_name_prefix = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.778746] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] profiler_otlp.service_name_prefix = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.778909] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] remote_debug.host = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.779079] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] remote_debug.port = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.779263] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.779427] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.779591] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.779790] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.779978] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.780158] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.780323] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.780485] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.780656] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.780829] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.780988] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.781172] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.781342] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.781514] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.781687] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.781855] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.782028] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.782208] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.782375] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.782555] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.782728] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.782893] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.783065] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.783234] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.783395] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.783562] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.783742] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.783909] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.784087] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.784256] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.ssl = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.784427] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.784594] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.784757] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.784948] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.785136] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.785300] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.785484] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.785656] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_notifications.retry = -1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.785834] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786010] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786190] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.auth_section = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786351] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.auth_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786509] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.cafile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786667] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.certfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786828] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.collect_timing = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.786984] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.connect_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.787156] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.connect_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.787314] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.endpoint_id = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.787470] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.endpoint_override = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.787630] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.insecure = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.787786] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.keyfile = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.787940] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.max_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.788105] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.min_version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.788265] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.region_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.788424] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.retriable_status_codes = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.788581] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.service_name = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.788737] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.service_type = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.788896] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.split_loggers = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.789062] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.status_code_retries = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.789222] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.status_code_retry_delay = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.789380] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.timeout = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.789537] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.valid_interfaces = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.789694] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_limit.version = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.789856] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_reports.file_event_handler = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790026] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790186] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] oslo_reports.log_dir = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790365] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790511] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790666] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790828] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.790988] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.791158] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.791325] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.791482] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_ovs_privileged.group = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.791640] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.791838] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.792037] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.792203] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] vif_plug_ovs_privileged.user = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.792371] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.792574] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.792756] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.792924] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.793108] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.793277] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.793443] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.793604] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.793783] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.793953] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.isolate_vif = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.794132] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.794301] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.794469] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.794638] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.794805] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] os_vif_ovs.per_port_bridge = False {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.794997] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] privsep_osbrick.capabilities = [21] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.795173] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] privsep_osbrick.group = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.795331] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] privsep_osbrick.helper_command = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.795494] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.795658] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.795817] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] privsep_osbrick.user = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.795987] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.796157] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] nova_sys_admin.group = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.796314] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] nova_sys_admin.helper_command = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.796475] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.796637] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.796842] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] nova_sys_admin.user = None {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 460.797013] env[63418]: DEBUG oslo_service.service [None req-a4649b49-5758-4ff9-b454-9026fc8642b3 None None] ******************************************************************************** {{(pid=63418) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 460.797507] env[63418]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 461.302053] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Getting list of instances from cluster (obj){ [ 461.302053] env[63418]: value = "domain-c8" [ 461.302053] env[63418]: _type = "ClusterComputeResource" [ 461.302053] env[63418]: } {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 461.302577] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88e0fea-d9ae-47fb-bfe6-6302681dd71a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.311355] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Got total of 0 instances {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 461.311880] env[63418]: WARNING nova.virt.vmwareapi.driver [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 461.312365] env[63418]: INFO nova.virt.node [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Generated node identity 6ac9de28-4c58-4fc2-8a3d-711092e3c63c [ 461.312655] env[63418]: INFO nova.virt.node [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Wrote node identity 6ac9de28-4c58-4fc2-8a3d-711092e3c63c to /opt/stack/data/n-cpu-1/compute_id [ 461.815350] env[63418]: WARNING nova.compute.manager [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Compute nodes ['6ac9de28-4c58-4fc2-8a3d-711092e3c63c'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 462.821487] env[63418]: INFO nova.compute.manager [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 463.830975] env[63418]: WARNING nova.compute.manager [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 463.831326] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 463.831488] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 463.831641] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 463.831810] env[63418]: DEBUG nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 463.832809] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb2ad36-d846-4026-a0eb-492ce3426e57 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 463.841597] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16584532-0c31-43da-af95-d32edfdaea62 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 463.857580] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10801438-564e-42eb-8271-5e37f48ab35a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 463.863853] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d36c24b-5604-4871-bfff-580659526b15 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 463.892912] env[63418]: DEBUG nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181546MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 463.893117] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 463.893333] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 464.399805] env[63418]: WARNING nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] No compute node record for cpu-1:6ac9de28-4c58-4fc2-8a3d-711092e3c63c: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 6ac9de28-4c58-4fc2-8a3d-711092e3c63c could not be found. [ 464.904082] env[63418]: INFO nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c [ 466.415647] env[63418]: DEBUG nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 466.416042] env[63418]: DEBUG nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 466.565706] env[63418]: INFO nova.scheduler.client.report [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] [req-18667199-9502-4090-b468-29b76f9cdfef] Created resource provider record via placement API for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 466.582611] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13891811-721d-40ba-bada-95eac99d9e01 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 466.590353] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a9f4b0-f35c-4eb8-9b48-2cd29487f5e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 466.621056] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d27ff7d-4717-404a-9230-9cefd2574999 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 466.628137] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318afc77-823d-4c83-b35b-3e91c1e75af0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 466.641121] env[63418]: DEBUG nova.compute.provider_tree [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 467.177664] env[63418]: DEBUG nova.scheduler.client.report [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 467.177915] env[63418]: DEBUG nova.compute.provider_tree [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 0 to 1 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 467.178090] env[63418]: DEBUG nova.compute.provider_tree [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 467.229296] env[63418]: DEBUG nova.compute.provider_tree [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 1 to 2 during operation: update_traits {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 467.733678] env[63418]: DEBUG nova.compute.resource_tracker [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 467.734084] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.841s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 467.734084] env[63418]: DEBUG nova.service [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Creating RPC server for service compute {{(pid=63418) start /opt/stack/nova/nova/service.py:186}} [ 467.748778] env[63418]: DEBUG nova.service [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] Join ServiceGroup membership for this service compute {{(pid=63418) start /opt/stack/nova/nova/service.py:203}} [ 467.748978] env[63418]: DEBUG nova.servicegroup.drivers.db [None req-e5f15296-ccfe-4254-9e6f-6d2c37ebfdd6 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63418) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 499.752570] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 500.261592] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Getting list of instances from cluster (obj){ [ 500.261592] env[63418]: value = "domain-c8" [ 500.261592] env[63418]: _type = "ClusterComputeResource" [ 500.261592] env[63418]: } {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 500.266272] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0f78dc-57a6-4fca-96d0-4a756342e3f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.276430] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Got total of 0 instances {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 500.276951] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 500.277398] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Getting list of instances from cluster (obj){ [ 500.277398] env[63418]: value = "domain-c8" [ 500.277398] env[63418]: _type = "ClusterComputeResource" [ 500.277398] env[63418]: } {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 500.278364] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f268fef5-2cf6-47f1-bbcf-05bd1675a03c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.292082] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Got total of 0 instances {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 503.328325] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquiring lock "8879ea02-d856-4233-a5e8-312701705119" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.328325] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Lock "8879ea02-d856-4233-a5e8-312701705119" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.837202] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 504.254683] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquiring lock "9122ae95-8678-48bd-9299-b217c77e7a0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.258023] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Lock "9122ae95-8678-48bd-9299-b217c77e7a0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.379534] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.379844] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.383090] env[63418]: INFO nova.compute.claims [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 504.736313] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.736734] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.765068] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 504.879916] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "ba67658a-668e-4fca-aefe-e838f7b05e2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.880194] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "ba67658a-668e-4fca-aefe-e838f7b05e2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.239053] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 505.292950] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.387700] env[63418]: DEBUG nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 505.516999] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317ed9c8-4607-4617-abb4-c7612c1de083 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.525525] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02866b2b-0fe2-485b-a9fb-15359f978cad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.555631] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0049a111-3e9f-4864-95d2-92ef4342e3a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.562937] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3168b1f-6c6e-4b3c-9f5d-7bfcf51e679a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.577407] env[63418]: DEBUG nova.compute.provider_tree [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 505.760896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.817495] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquiring lock "8fbc105f-34f9-4ece-9e74-e473f310221a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.818542] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Lock "8fbc105f-34f9-4ece-9e74-e473f310221a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.916993] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.080658] env[63418]: DEBUG nova.scheduler.client.report [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 506.320045] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 506.589287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.209s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 506.589974] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 506.593564] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.300s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.597746] env[63418]: INFO nova.compute.claims [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 506.858638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.102236] env[63418]: DEBUG nova.compute.utils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 507.109024] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 507.109024] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 507.518919] env[63418]: DEBUG nova.policy [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70e7a7b14ac446c4bd6a8eb6c63707ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f5d225c9fa94572b4d17878c98c2408', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 507.612715] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 507.750780] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e527a32-6fcc-4071-ad7a-794cb30bed8e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.759551] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fec518-0af0-476d-b2c2-9f9aa21bf48f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.796615] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4b49a8-3cf9-4950-b58e-6136255918b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.804646] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdff75ee-34fb-432a-a618-c2be3d406c12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.818764] env[63418]: DEBUG nova.compute.provider_tree [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 508.095249] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Successfully created port: 231b1082-0989-4ab9-a2e1-60ae042d8927 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 508.322409] env[63418]: DEBUG nova.scheduler.client.report [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 508.625421] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 508.661866] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 508.662571] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 508.662571] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 508.662799] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 508.663308] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 508.664142] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 508.664142] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 508.664142] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 508.664555] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 508.664737] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 508.664911] env[63418]: DEBUG nova.virt.hardware [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 508.666175] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60ee82b-6738-4f83-8c89-804abb18fe83 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.681076] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f885f7bd-917f-4bd2-9766-8de5afe392ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.712530] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c5d0c7-0303-45ec-968c-6e7f02b883a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.830414] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 508.831693] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 508.835507] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.075s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.839250] env[63418]: INFO nova.compute.claims [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 509.342082] env[63418]: DEBUG nova.compute.utils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 509.344155] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 509.345157] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 509.511325] env[63418]: DEBUG nova.policy [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a52750eebe7d43fe9069f424924b251a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'efd17976b1f24e1a8e72250fd19ef13c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 509.848978] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 509.958157] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7411b318-d3f9-4291-87b8-113b56219967 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.969113] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae11c6d8-faf2-4157-8036-a9e022e29225 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.003385] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d298a9-a656-4a1e-8539-c97c8362a39c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.012125] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a9510b-84a8-46e9-82a3-fdd1b87756a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.026915] env[63418]: DEBUG nova.compute.provider_tree [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 510.534296] env[63418]: DEBUG nova.scheduler.client.report [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 510.854570] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Successfully created port: acde18ca-d2f2-4f91-ab71-cf0bdec52b55 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 510.867808] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 510.904526] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 510.904526] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 510.904526] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 510.904787] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 510.904787] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 510.905089] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 510.905393] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 510.905621] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 510.905874] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 510.906238] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 510.906488] env[63418]: DEBUG nova.virt.hardware [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 510.907404] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6847db98-c230-4d2f-8e21-a78e0f5dee14 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.917370] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325901b3-3117-43e5-8eb6-34b9feff2a20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.043811] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 511.044372] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 511.048294] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.131s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.050705] env[63418]: INFO nova.compute.claims [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 511.555505] env[63418]: DEBUG nova.compute.utils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 511.558767] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 511.559555] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 511.601943] env[63418]: ERROR nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 511.601943] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 511.601943] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.601943] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.601943] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.601943] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.601943] env[63418]: ERROR nova.compute.manager raise self.value [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.601943] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 511.601943] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.601943] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 511.602945] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.602945] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 511.602945] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 511.602945] env[63418]: ERROR nova.compute.manager [ 511.602945] env[63418]: Traceback (most recent call last): [ 511.602945] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 511.602945] env[63418]: listener.cb(fileno) [ 511.602945] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.602945] env[63418]: result = function(*args, **kwargs) [ 511.602945] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 511.602945] env[63418]: return func(*args, **kwargs) [ 511.602945] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 511.602945] env[63418]: raise e [ 511.602945] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 511.602945] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 511.602945] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.602945] env[63418]: created_port_ids = self._update_ports_for_instance( [ 511.602945] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.602945] env[63418]: with excutils.save_and_reraise_exception(): [ 511.602945] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.602945] env[63418]: self.force_reraise() [ 511.602945] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.602945] env[63418]: raise self.value [ 511.602945] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.602945] env[63418]: updated_port = self._update_port( [ 511.602945] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.602945] env[63418]: _ensure_no_port_binding_failure(port) [ 511.602945] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.602945] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 511.603721] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 511.603721] env[63418]: Removing descriptor: 15 [ 511.604079] env[63418]: ERROR nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] Traceback (most recent call last): [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] yield resources [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self.driver.spawn(context, instance, image_meta, [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self._vmops.spawn(context, instance, image_meta, injected_files, [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] vm_ref = self.build_virtual_machine(instance, [ 511.604079] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] vif_infos = vmwarevif.get_vif_info(self._session, [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] for vif in network_info: [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return self._sync_wrapper(fn, *args, **kwargs) [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self.wait() [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self[:] = self._gt.wait() [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return self._exit_event.wait() [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 511.604891] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] result = hub.switch() [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return self.greenlet.switch() [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] result = function(*args, **kwargs) [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return func(*args, **kwargs) [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] raise e [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] nwinfo = self.network_api.allocate_for_instance( [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] created_port_ids = self._update_ports_for_instance( [ 511.605304] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] with excutils.save_and_reraise_exception(): [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self.force_reraise() [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] raise self.value [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] updated_port = self._update_port( [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] _ensure_no_port_binding_failure(port) [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] raise exception.PortBindingFailed(port_id=port['id']) [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 511.605646] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] [ 511.605977] env[63418]: INFO nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Terminating instance [ 511.783573] env[63418]: DEBUG nova.policy [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83c6ebfffc204cb8ad6e61c5f4569017', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3c5f878ad3b4aae9da1afcdbdba4ec8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 512.060211] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 512.111959] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquiring lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.113751] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquired lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.113751] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 512.257222] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecded8e9-83f4-471c-bd40-5f3d12675347 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.271033] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2aab55a-d2a2-43ef-bcc2-6afb73604450 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.308379] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eda14ba-09bf-46a5-9792-de66540186c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.320608] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929d298b-474c-4827-a560-a61423c36bd0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.336561] env[63418]: DEBUG nova.compute.provider_tree [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.645591] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "7e43d259-f361-43d8-8f03-72b303680478" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.645591] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "7e43d259-f361-43d8-8f03-72b303680478" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.677984] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 512.846188] env[63418]: DEBUG nova.scheduler.client.report [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 512.999093] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.080359] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 513.115027] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 513.115563] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 513.115563] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 513.115667] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 513.115913] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 513.115913] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 513.116217] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 513.116297] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 513.116475] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 513.116630] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 513.116791] env[63418]: DEBUG nova.virt.hardware [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 513.118024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cffd45a-7545-402c-b0ce-0d2a25b7184b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.127379] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a257d655-5820-4847-a8ff-7c3778c37847 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.147375] env[63418]: DEBUG nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 513.162232] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquiring lock "9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.162690] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Lock "9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.246635] env[63418]: DEBUG nova.compute.manager [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] [instance: 8879ea02-d856-4233-a5e8-312701705119] Received event network-changed-231b1082-0989-4ab9-a2e1-60ae042d8927 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 513.246783] env[63418]: DEBUG nova.compute.manager [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] [instance: 8879ea02-d856-4233-a5e8-312701705119] Refreshing instance network info cache due to event network-changed-231b1082-0989-4ab9-a2e1-60ae042d8927. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 513.246945] env[63418]: DEBUG oslo_concurrency.lockutils [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] Acquiring lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.279847] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Successfully created port: 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 513.350916] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.351494] env[63418]: DEBUG nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 513.356200] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.496s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.356200] env[63418]: INFO nova.compute.claims [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.501328] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Releasing lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.501693] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 513.501901] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 513.502200] env[63418]: DEBUG oslo_concurrency.lockutils [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] Acquired lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.502375] env[63418]: DEBUG nova.network.neutron [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] [instance: 8879ea02-d856-4233-a5e8-312701705119] Refreshing network info cache for port 231b1082-0989-4ab9-a2e1-60ae042d8927 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 513.503478] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed338367-31f3-437d-bfe1-7d13544fdec2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.524561] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfe3c96-9be5-4790-9387-2955c00f83c6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.558897] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8879ea02-d856-4233-a5e8-312701705119 could not be found. [ 513.558897] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 513.559097] env[63418]: INFO nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Took 0.06 seconds to destroy the instance on the hypervisor. [ 513.559556] env[63418]: DEBUG oslo.service.loopingcall [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 513.559915] env[63418]: DEBUG nova.compute.manager [-] [instance: 8879ea02-d856-4233-a5e8-312701705119] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 513.560304] env[63418]: DEBUG nova.network.neutron [-] [instance: 8879ea02-d856-4233-a5e8-312701705119] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 513.666818] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 513.681160] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.802352] env[63418]: DEBUG nova.network.neutron [-] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.861038] env[63418]: DEBUG nova.compute.utils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.861038] env[63418]: DEBUG nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 514.067472] env[63418]: DEBUG nova.network.neutron [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 514.188249] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.306177] env[63418]: DEBUG nova.network.neutron [-] [instance: 8879ea02-d856-4233-a5e8-312701705119] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.352757] env[63418]: DEBUG nova.network.neutron [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] [instance: 8879ea02-d856-4233-a5e8-312701705119] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.365856] env[63418]: DEBUG nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 514.556988] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d48e6ed-34e8-4a0e-a9f5-93c1dcb66c6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.565200] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd473f5-3a7d-464f-b3a1-57501bc7cf6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.609183] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1351010b-6271-4830-b2fd-294d2f2bf9ff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.620523] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685e2a9b-816b-40b2-ac49-630696ce1e78 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.637420] env[63418]: DEBUG nova.compute.provider_tree [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 514.808982] env[63418]: INFO nova.compute.manager [-] [instance: 8879ea02-d856-4233-a5e8-312701705119] Took 1.25 seconds to deallocate network for instance. [ 514.814863] env[63418]: DEBUG nova.compute.claims [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 514.815505] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.855864] env[63418]: DEBUG oslo_concurrency.lockutils [req-180e6860-d095-4459-8093-7c13c18d3edc req-f8e4392d-6546-4687-a3ac-907caade3046 service nova] Releasing lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.141170] env[63418]: DEBUG nova.scheduler.client.report [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 515.377865] env[63418]: DEBUG nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 515.407276] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.407276] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.407276] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.407276] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.407767] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.408601] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.408845] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.409266] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.409266] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.409378] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.409600] env[63418]: DEBUG nova.virt.hardware [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.410626] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55c6091-61f2-411b-86c5-f84ee7e4fcc1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.419740] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e4d086-7b70-4f46-b136-e4045196d900 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.434100] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 515.443387] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.443725] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d181406e-d1a6-453a-b97d-e203101be087 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.456844] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Created folder: OpenStack in parent group-v4. [ 515.457097] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating folder: Project (36d63ad5fcea4450adc6c714bfcbb800). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.457294] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-37184eb1-12c0-4687-ad83-215c81878915 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.468404] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Created folder: Project (36d63ad5fcea4450adc6c714bfcbb800) in parent group-v268354. [ 515.468404] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating folder: Instances. Parent ref: group-v268355. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.468404] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a89f2fa-8871-403d-99d7-770a129c1888 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.477190] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Created folder: Instances in parent group-v268355. [ 515.477452] env[63418]: DEBUG oslo.service.loopingcall [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 515.477646] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 515.477857] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6e07d2e-fea7-4055-8491-29ee9c608daa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.496905] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 515.496905] env[63418]: value = "task-1244597" [ 515.496905] env[63418]: _type = "Task" [ 515.496905] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.508474] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244597, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.641588] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "f185a348-e91d-48d1-970b-473cc253cfdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.641823] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "f185a348-e91d-48d1-970b-473cc253cfdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.646062] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.646529] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 515.649754] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.969s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.653283] env[63418]: INFO nova.compute.claims [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.728083] env[63418]: ERROR nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 515.728083] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 515.728083] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 515.728083] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 515.728083] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.728083] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.728083] env[63418]: ERROR nova.compute.manager raise self.value [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 515.728083] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 515.728083] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.728083] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 515.730888] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.730888] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 515.730888] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 515.730888] env[63418]: ERROR nova.compute.manager [ 515.730888] env[63418]: Traceback (most recent call last): [ 515.730888] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 515.730888] env[63418]: listener.cb(fileno) [ 515.730888] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.730888] env[63418]: result = function(*args, **kwargs) [ 515.730888] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.730888] env[63418]: return func(*args, **kwargs) [ 515.730888] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 515.730888] env[63418]: raise e [ 515.730888] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 515.730888] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 515.730888] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 515.730888] env[63418]: created_port_ids = self._update_ports_for_instance( [ 515.730888] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 515.730888] env[63418]: with excutils.save_and_reraise_exception(): [ 515.730888] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.730888] env[63418]: self.force_reraise() [ 515.730888] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.730888] env[63418]: raise self.value [ 515.730888] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 515.730888] env[63418]: updated_port = self._update_port( [ 515.730888] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.730888] env[63418]: _ensure_no_port_binding_failure(port) [ 515.730888] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.730888] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 515.733858] env[63418]: nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 515.733858] env[63418]: Removing descriptor: 16 [ 515.733858] env[63418]: ERROR nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Traceback (most recent call last): [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] yield resources [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self.driver.spawn(context, instance, image_meta, [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 515.733858] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] vm_ref = self.build_virtual_machine(instance, [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] for vif in network_info: [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return self._sync_wrapper(fn, *args, **kwargs) [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self.wait() [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self[:] = self._gt.wait() [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return self._exit_event.wait() [ 515.734298] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] result = hub.switch() [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return self.greenlet.switch() [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] result = function(*args, **kwargs) [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return func(*args, **kwargs) [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] raise e [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] nwinfo = self.network_api.allocate_for_instance( [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 515.734671] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] created_port_ids = self._update_ports_for_instance( [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] with excutils.save_and_reraise_exception(): [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self.force_reraise() [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] raise self.value [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] updated_port = self._update_port( [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] _ensure_no_port_binding_failure(port) [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.735802] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] raise exception.PortBindingFailed(port_id=port['id']) [ 515.736515] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 515.736515] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] [ 515.736515] env[63418]: INFO nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Terminating instance [ 516.009505] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244597, 'name': CreateVM_Task, 'duration_secs': 0.345146} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.009684] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 516.010730] env[63418]: DEBUG oslo_vmware.service [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740a5f40-a232-4eaa-b936-22c61babd0b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.020064] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.020501] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.021371] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 516.021979] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d901a1a3-f524-4273-bf89-4753cf738208 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.027038] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 516.027038] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]524f2600-e3ee-6446-ed70-309eb0aa52b7" [ 516.027038] env[63418]: _type = "Task" [ 516.027038] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.037969] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524f2600-e3ee-6446-ed70-309eb0aa52b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.084792] env[63418]: ERROR nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 516.084792] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 516.084792] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 516.084792] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 516.084792] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 516.084792] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 516.084792] env[63418]: ERROR nova.compute.manager raise self.value [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 516.084792] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 516.084792] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 516.084792] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 516.085325] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 516.085325] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 516.085325] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 516.085325] env[63418]: ERROR nova.compute.manager [ 516.085325] env[63418]: Traceback (most recent call last): [ 516.085325] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 516.085325] env[63418]: listener.cb(fileno) [ 516.085325] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 516.085325] env[63418]: result = function(*args, **kwargs) [ 516.085325] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 516.085325] env[63418]: return func(*args, **kwargs) [ 516.085325] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 516.085325] env[63418]: raise e [ 516.085325] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 516.085325] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 516.085325] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 516.085325] env[63418]: created_port_ids = self._update_ports_for_instance( [ 516.085325] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 516.085325] env[63418]: with excutils.save_and_reraise_exception(): [ 516.085325] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 516.085325] env[63418]: self.force_reraise() [ 516.085325] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 516.085325] env[63418]: raise self.value [ 516.085325] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 516.085325] env[63418]: updated_port = self._update_port( [ 516.085325] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 516.085325] env[63418]: _ensure_no_port_binding_failure(port) [ 516.085325] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 516.085325] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 516.086149] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 516.086149] env[63418]: Removing descriptor: 17 [ 516.086149] env[63418]: ERROR nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Traceback (most recent call last): [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] yield resources [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self.driver.spawn(context, instance, image_meta, [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 516.086149] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] vm_ref = self.build_virtual_machine(instance, [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] for vif in network_info: [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return self._sync_wrapper(fn, *args, **kwargs) [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self.wait() [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self[:] = self._gt.wait() [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return self._exit_event.wait() [ 516.086576] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] result = hub.switch() [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return self.greenlet.switch() [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] result = function(*args, **kwargs) [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return func(*args, **kwargs) [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] raise e [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] nwinfo = self.network_api.allocate_for_instance( [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 516.086925] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] created_port_ids = self._update_ports_for_instance( [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] with excutils.save_and_reraise_exception(): [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self.force_reraise() [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] raise self.value [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] updated_port = self._update_port( [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] _ensure_no_port_binding_failure(port) [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 516.087360] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] raise exception.PortBindingFailed(port_id=port['id']) [ 516.087708] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 516.087708] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] [ 516.087708] env[63418]: INFO nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Terminating instance [ 516.146627] env[63418]: DEBUG nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 516.157098] env[63418]: DEBUG nova.compute.utils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 516.161863] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 516.161863] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 516.212305] env[63418]: DEBUG nova.policy [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd716120538564d34984002051ffcdf97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fac5fb783c549969000a6a930897712', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 516.224988] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.225125] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.225279] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 516.225346] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Rebuilding the list of instances to heal {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10282}} [ 516.239595] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquiring lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.239595] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquired lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.239595] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 516.547675] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.547953] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 516.549057] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.549057] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.549057] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 516.549305] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce80e04c-75cc-4638-9afa-2d8df414e18c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.571348] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 516.571348] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 516.571811] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb126e0-dcca-4d3a-a6f9-bd6a7e89c0d5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.586211] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cee39454-f3d8-4e34-acd2-5caa445b3d80 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.597848] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.598058] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquired lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.598333] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 516.599827] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 516.599827] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f11703-9280-d39e-567b-b2d0b0b8c8f9" [ 516.599827] env[63418]: _type = "Task" [ 516.599827] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.608169] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f11703-9280-d39e-567b-b2d0b0b8c8f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.662345] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 516.686249] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.734291] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 8879ea02-d856-4233-a5e8-312701705119] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 516.734537] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 516.734797] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 516.734895] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 516.736292] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 516.736292] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 516.736292] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Didn't find any instances for network info cache update. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10364}} [ 516.740212] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.741670] env[63418]: DEBUG nova.compute.manager [req-3fb7a7d8-3671-4d93-8cae-abaac7695653 req-66cba392-1e6a-4eeb-b7f6-d15c6a51285e service nova] [instance: 8879ea02-d856-4233-a5e8-312701705119] Received event network-vif-deleted-231b1082-0989-4ab9-a2e1-60ae042d8927 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 516.744687] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.749114] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.750264] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.753687] env[63418]: DEBUG nova.compute.manager [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Received event network-changed-acde18ca-d2f2-4f91-ab71-cf0bdec52b55 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 516.754424] env[63418]: DEBUG nova.compute.manager [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Refreshing instance network info cache due to event network-changed-acde18ca-d2f2-4f91-ab71-cf0bdec52b55. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 516.754654] env[63418]: DEBUG oslo_concurrency.lockutils [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] Acquiring lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.755361] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.756232] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.756472] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 516.756885] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 516.813367] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 516.827180] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Successfully created port: 0bccf0ff-e310-44de-a851-c3dffcfd8e8a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.865637] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ac9e26-5460-465c-90d3-5db988e60992 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.874059] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48870dbf-5f13-449b-88a2-b66c2c560162 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.923874] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa0a818-6e93-4965-9a1c-0919cb4084ad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.931285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4db0898-8c1d-4a71-970f-baa19ca5dc0c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.945448] env[63418]: DEBUG nova.compute.provider_tree [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.064557] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.118847] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 517.119346] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating directory with path [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 517.119714] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fba95738-3655-40bd-a006-5371d037278f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.141491] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 517.143616] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "d255e7af-b50d-4b30-8bec-9e2ee2de2c80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.143871] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "d255e7af-b50d-4b30-8bec-9e2ee2de2c80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.161600] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Created directory with path [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 517.161600] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Fetch image to [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 517.161718] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Downloading image file data c0dccf29-5e49-4a1e-b51f-d46e566b4772 to [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk on the data store datastore2 {{(pid=63418) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 517.163410] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f53c53b-4c70-4793-9a33-327152633459 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.174446] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c92d44-d3ad-457a-913f-83214cca5e40 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.192605] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86eb5235-2ca1-441d-89e1-f031fe5612bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.241636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867d5fd1-b02d-438a-95ff-97292df0598c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.251777] env[63418]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-9dac7533-b5cc-4a62-b3e5-36f85b6a8aa7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.259496] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.285209] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Downloading image file data c0dccf29-5e49-4a1e-b51f-d46e566b4772 to the data store datastore2 {{(pid=63418) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 517.334593] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.379609] env[63418]: DEBUG oslo_vmware.rw_handles [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 517.452793] env[63418]: DEBUG nova.scheduler.client.report [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 517.571580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Releasing lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.571580] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 517.571580] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 517.572130] env[63418]: DEBUG oslo_concurrency.lockutils [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] Acquired lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.572130] env[63418]: DEBUG nova.network.neutron [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Refreshing network info cache for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 517.573109] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a5c53e0-949a-40a1-8ccb-54af90122371 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.595782] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249211fa-a4ce-42ef-8d8e-27da2fff5c68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.622717] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9122ae95-8678-48bd-9299-b217c77e7a0d could not be found. [ 517.622976] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 517.623184] env[63418]: INFO nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 517.623428] env[63418]: DEBUG oslo.service.loopingcall [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 517.625662] env[63418]: DEBUG nova.compute.manager [-] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 517.625747] env[63418]: DEBUG nova.network.neutron [-] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 517.651104] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 517.657826] env[63418]: DEBUG nova.network.neutron [-] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 517.683426] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 517.723811] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 517.723811] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 517.723811] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 517.724124] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 517.724124] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 517.724124] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 517.724124] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 517.724124] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 517.724385] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 517.724385] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 517.724385] env[63418]: DEBUG nova.virt.hardware [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 517.724385] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85913a8f-9d7c-4efb-a9e2-29433f7508b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.737199] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df45fb2e-6b13-4f47-9b5a-93d27bcf8012 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.838881] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Releasing lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.840411] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 517.840411] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 517.842382] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4654da0-6bb6-4a0d-b618-72e8b4c29488 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.855102] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68535f43-0b8e-4f8f-80de-85d0e8920db7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.887370] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf could not be found. [ 517.888046] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 517.888046] env[63418]: INFO nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 517.888370] env[63418]: DEBUG oslo.service.loopingcall [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 517.890810] env[63418]: DEBUG nova.compute.manager [-] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 517.891189] env[63418]: DEBUG nova.network.neutron [-] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 517.937133] env[63418]: DEBUG nova.network.neutron [-] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 517.961986] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.962534] env[63418]: DEBUG nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 517.972976] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.785s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.978423] env[63418]: INFO nova.compute.claims [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.050078] env[63418]: DEBUG oslo_vmware.rw_handles [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 518.050460] env[63418]: DEBUG oslo_vmware.rw_handles [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 518.162255] env[63418]: DEBUG nova.network.neutron [-] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.179141] env[63418]: DEBUG nova.network.neutron [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 518.181595] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.196154] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Downloaded image file data c0dccf29-5e49-4a1e-b51f-d46e566b4772 to vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk on the data store datastore2 {{(pid=63418) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 518.198421] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 518.202021] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Copying Virtual Disk [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk to [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 518.202021] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f59f088-5318-481c-a3cc-1a652fa7d8ca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.209077] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 518.209077] env[63418]: value = "task-1244598" [ 518.209077] env[63418]: _type = "Task" [ 518.209077] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.221018] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244598, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.441453] env[63418]: DEBUG nova.network.neutron [-] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.465166] env[63418]: DEBUG nova.network.neutron [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.477505] env[63418]: DEBUG nova.compute.utils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.477505] env[63418]: DEBUG nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 518.667585] env[63418]: INFO nova.compute.manager [-] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Took 1.04 seconds to deallocate network for instance. [ 518.672443] env[63418]: DEBUG nova.compute.claims [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 518.672443] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.723242] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244598, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.944576] env[63418]: INFO nova.compute.manager [-] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Took 1.05 seconds to deallocate network for instance. [ 518.950122] env[63418]: DEBUG nova.compute.claims [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 518.951034] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.967390] env[63418]: DEBUG oslo_concurrency.lockutils [req-dea33277-d444-43fd-bd5a-4930b837df51 req-ad914b6e-efb3-4020-af66-d905783c7d16 service nova] Releasing lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 518.978724] env[63418]: DEBUG nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 519.185458] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56ace9b-ee59-4456-a2b7-1c70e22990da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.198273] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "781262a7-56ba-466c-86b3-1551905c08bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.198504] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "781262a7-56ba-466c-86b3-1551905c08bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.202034] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525d1220-f2c3-4003-90d8-db6dbffcda46 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.245216] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886f1f28-501d-4517-bb44-2dd9f158aefc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.255478] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244598, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.722918} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.257714] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Copied Virtual Disk [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk to [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 519.258081] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleting the datastore file [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 519.259131] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21858c8e-5c8d-4223-a240-0bef55e288c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.266595] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7903ad76-bcb2-48e7-bfae-9fbde81d7d55 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.272471] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 519.272471] env[63418]: value = "task-1244599" [ 519.272471] env[63418]: _type = "Task" [ 519.272471] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.285828] env[63418]: DEBUG nova.compute.provider_tree [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.291958] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.637526] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "01deeb21-0e27-497f-9b85-c85949a3533d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.638065] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "01deeb21-0e27-497f-9b85-c85949a3533d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.706667] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 519.785646] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023883} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.785971] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 519.786233] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Moving file from [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738/c0dccf29-5e49-4a1e-b51f-d46e566b4772 to [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772. {{(pid=63418) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 519.786485] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a80a4d60-ff6d-4e72-afbb-6ed5cf386608 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.789505] env[63418]: DEBUG nova.scheduler.client.report [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 519.798555] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 519.798555] env[63418]: value = "task-1244600" [ 519.798555] env[63418]: _type = "Task" [ 519.798555] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.808190] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244600, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.863821] env[63418]: DEBUG nova.compute.manager [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Received event network-changed-76f691d8-a4d1-409c-9ff2-b21e7dea3ac1 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 519.863971] env[63418]: DEBUG nova.compute.manager [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Refreshing instance network info cache due to event network-changed-76f691d8-a4d1-409c-9ff2-b21e7dea3ac1. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 519.864193] env[63418]: DEBUG oslo_concurrency.lockutils [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] Acquiring lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.864294] env[63418]: DEBUG oslo_concurrency.lockutils [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] Acquired lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.864463] env[63418]: DEBUG nova.network.neutron [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Refreshing network info cache for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 519.876140] env[63418]: DEBUG nova.compute.manager [req-0ed7bb34-40f0-476f-802e-6f6f76f86089 req-db75f17a-2f08-4f84-8eb6-d060d980f95a service nova] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Received event network-vif-deleted-acde18ca-d2f2-4f91-ab71-cf0bdec52b55 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 519.990510] env[63418]: DEBUG nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 520.021510] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.021756] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.021909] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.022094] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.022254] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.022371] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.022724] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.022928] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.023115] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.023275] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.023438] env[63418]: DEBUG nova.virt.hardware [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.024360] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f80d1f-9756-4154-ab00-00b3497bda89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.033061] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e312f546-f6a8-4804-841b-b8e55a4192af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.052547] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 520.057808] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating folder: Project (bf550eb38ef8440297dbd0439c2b3eb8). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 520.058196] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3bc9929-9701-4425-bbda-c37f6c1d0a20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.068420] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Created folder: Project (bf550eb38ef8440297dbd0439c2b3eb8) in parent group-v268354. [ 520.068972] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating folder: Instances. Parent ref: group-v268358. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 520.068972] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c3e2dbb-632c-450a-8e7f-bbb8080bb6ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.082016] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Created folder: Instances in parent group-v268358. [ 520.082016] env[63418]: DEBUG oslo.service.loopingcall [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.082016] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 520.082016] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e77c539c-77a2-4cf0-9c3f-8f5a4ddb661f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.098412] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.098583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.102818] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 520.102818] env[63418]: value = "task-1244603" [ 520.102818] env[63418]: _type = "Task" [ 520.102818] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.110868] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244603, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.241939] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.295990] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.296596] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 520.301131] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.486s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.317310] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244600, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.045107} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.317403] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] File moved {{(pid=63418) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 520.317612] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Cleaning up location [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 520.317820] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleting the datastore file [datastore2] vmware_temp/2acb8bbc-11cb-4cbb-b2ed-6fd572935738 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 520.318106] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5c6d487-61ec-4f83-aaad-7031053d9b5b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.328185] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 520.328185] env[63418]: value = "task-1244604" [ 520.328185] env[63418]: _type = "Task" [ 520.328185] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.341944] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.403466] env[63418]: DEBUG nova.network.neutron [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.569208] env[63418]: DEBUG nova.network.neutron [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.612888] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244603, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.807439] env[63418]: DEBUG nova.compute.utils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.807980] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 520.808177] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 520.844567] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024444} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.848604] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 520.849836] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54da4b5e-a0ae-4278-8b1b-5400b7112fb3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.861049] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 520.861049] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d5904b-8805-fcb7-32c8-3a32ab85ee30" [ 520.861049] env[63418]: _type = "Task" [ 520.861049] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.869023] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d5904b-8805-fcb7-32c8-3a32ab85ee30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.015101] env[63418]: DEBUG nova.policy [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfb16a1f831f4f99afa8fdac55953937', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ff1f143cbe544ba8f3156142c0e03d4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.073698] env[63418]: DEBUG oslo_concurrency.lockutils [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] Releasing lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.073698] env[63418]: DEBUG nova.compute.manager [req-13a0570e-e97a-4e85-9434-2269e8a412cf req-4a1df0dc-a55c-40cb-8fea-62019f7ccc3b service nova] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Received event network-vif-deleted-76f691d8-a4d1-409c-9ff2-b21e7dea3ac1 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 521.086686] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970b0fd9-43fc-4617-b5d1-584a1a8af03a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.100166] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f129e669-f5eb-40ea-8db0-c4d2d0329eb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.116018] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244603, 'name': CreateVM_Task, 'duration_secs': 0.598977} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.160798] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 521.166146] env[63418]: DEBUG oslo_vmware.service [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0322962-bce0-408c-9907-b85b826d3881 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.172154] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabcc974-64c5-4565-9098-e8e7fba23c87 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.177476] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.179016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.179016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 521.180203] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b11e1e54-88bf-4372-a328-3e4bd48ff0d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.182887] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ccb693-89cb-4dae-a10f-367b38c28979 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.192806] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 521.192806] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5264944e-c818-2dcd-cf4c-7f4db3c39058" [ 521.192806] env[63418]: _type = "Task" [ 521.192806] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.199460] env[63418]: DEBUG nova.compute.provider_tree [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.209560] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.211514] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 521.211514] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.211514] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.211514] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 521.211514] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1eca3aad-6727-49ab-a14a-5a998e50f810 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.219816] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 521.220081] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 521.220878] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb4b44-2e3f-4a64-803a-1779413a8172 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.227659] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02879ee3-778d-485c-a6e0-f1c7051ea656 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.233482] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 521.233482] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aba2c6-bce6-c0b9-1417-88def16fb991" [ 521.233482] env[63418]: _type = "Task" [ 521.233482] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.243405] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aba2c6-bce6-c0b9-1417-88def16fb991, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.311874] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 521.371996] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d5904b-8805-fcb7-32c8-3a32ab85ee30, 'name': SearchDatastore_Task, 'duration_secs': 0.009546} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.373153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.373492] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 521.373660] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29586c0d-bee3-40c2-b6ca-a2d1610b590a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.386205] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 521.386205] env[63418]: value = "task-1244605" [ 521.386205] env[63418]: _type = "Task" [ 521.386205] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.395593] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.705998] env[63418]: DEBUG nova.scheduler.client.report [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 521.725601] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "a8c7f192-4672-43cc-8c38-0c33ce633765" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.726921] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "a8c7f192-4672-43cc-8c38-0c33ce633765" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.754238] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 521.754585] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating directory with path [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 521.758342] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6827dbb7-7708-4f7b-9cac-78a19b939c11 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.782414] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Created directory with path [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 521.782963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Fetch image to [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 521.783308] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Downloading image file data c0dccf29-5e49-4a1e-b51f-d46e566b4772 to [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk on the data store datastore1 {{(pid=63418) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 521.784089] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90184987-a38a-4dc1-84ae-62da4a41d677 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.795298] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1805fa7-80b0-4c86-b580-7b0489088150 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.806451] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6750bd2d-f2aa-4d99-aa88-32e337f74bb5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.852267] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a255cf7-dc25-4561-b5e0-336927ee8fb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.860401] env[63418]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-ba135cb2-aba9-4b9c-93da-7c8c5f056b29 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.894564] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Downloading image file data c0dccf29-5e49-4a1e-b51f-d46e566b4772 to the data store datastore1 {{(pid=63418) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 521.903409] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502059} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.903771] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 521.904211] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 521.904211] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-faf2cccd-5a51-4093-b1e3-76f69e04ade7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.911189] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 521.911189] env[63418]: value = "task-1244606" [ 521.911189] env[63418]: _type = "Task" [ 521.911189] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.921014] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244606, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.936198] env[63418]: ERROR nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 521.936198] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 521.936198] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.936198] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.936198] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.936198] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.936198] env[63418]: ERROR nova.compute.manager raise self.value [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.936198] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 521.936198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.936198] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 521.936643] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.936643] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 521.936643] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 521.936643] env[63418]: ERROR nova.compute.manager [ 521.936643] env[63418]: Traceback (most recent call last): [ 521.936643] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 521.936643] env[63418]: listener.cb(fileno) [ 521.936643] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.936643] env[63418]: result = function(*args, **kwargs) [ 521.936643] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 521.936643] env[63418]: return func(*args, **kwargs) [ 521.936643] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 521.936643] env[63418]: raise e [ 521.936643] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 521.936643] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 521.936643] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.936643] env[63418]: created_port_ids = self._update_ports_for_instance( [ 521.936643] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.936643] env[63418]: with excutils.save_and_reraise_exception(): [ 521.936643] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.936643] env[63418]: self.force_reraise() [ 521.936643] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.936643] env[63418]: raise self.value [ 521.936643] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.936643] env[63418]: updated_port = self._update_port( [ 521.936643] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.936643] env[63418]: _ensure_no_port_binding_failure(port) [ 521.936643] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.936643] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 521.938090] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 521.938090] env[63418]: Removing descriptor: 17 [ 521.938090] env[63418]: ERROR nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Traceback (most recent call last): [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] yield resources [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self.driver.spawn(context, instance, image_meta, [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 521.938090] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] vm_ref = self.build_virtual_machine(instance, [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] vif_infos = vmwarevif.get_vif_info(self._session, [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] for vif in network_info: [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return self._sync_wrapper(fn, *args, **kwargs) [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self.wait() [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self[:] = self._gt.wait() [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return self._exit_event.wait() [ 521.938489] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] result = hub.switch() [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return self.greenlet.switch() [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] result = function(*args, **kwargs) [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return func(*args, **kwargs) [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] raise e [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] nwinfo = self.network_api.allocate_for_instance( [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.938895] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] created_port_ids = self._update_ports_for_instance( [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] with excutils.save_and_reraise_exception(): [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self.force_reraise() [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] raise self.value [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] updated_port = self._update_port( [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] _ensure_no_port_binding_failure(port) [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.939255] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] raise exception.PortBindingFailed(port_id=port['id']) [ 521.939956] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 521.939956] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] [ 521.939956] env[63418]: INFO nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Terminating instance [ 521.969650] env[63418]: DEBUG oslo_vmware.rw_handles [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 522.212271] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.213169] env[63418]: ERROR nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] Traceback (most recent call last): [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self.driver.spawn(context, instance, image_meta, [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] vm_ref = self.build_virtual_machine(instance, [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.213169] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] for vif in network_info: [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return self._sync_wrapper(fn, *args, **kwargs) [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self.wait() [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self[:] = self._gt.wait() [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return self._exit_event.wait() [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] result = hub.switch() [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.213480] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return self.greenlet.switch() [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] result = function(*args, **kwargs) [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] return func(*args, **kwargs) [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] raise e [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] nwinfo = self.network_api.allocate_for_instance( [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] created_port_ids = self._update_ports_for_instance( [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] with excutils.save_and_reraise_exception(): [ 522.213955] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] self.force_reraise() [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] raise self.value [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] updated_port = self._update_port( [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] _ensure_no_port_binding_failure(port) [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] raise exception.PortBindingFailed(port_id=port['id']) [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] nova.exception.PortBindingFailed: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. [ 522.214416] env[63418]: ERROR nova.compute.manager [instance: 8879ea02-d856-4233-a5e8-312701705119] [ 522.214796] env[63418]: DEBUG nova.compute.utils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 522.216419] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.530s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.218976] env[63418]: INFO nova.compute.claims [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.226098] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Successfully created port: 1d10ead4-73ff-4946-bf33-e3e9a105b862 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.228936] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Build of instance 8879ea02-d856-4233-a5e8-312701705119 was re-scheduled: Binding failed for port 231b1082-0989-4ab9-a2e1-60ae042d8927, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 522.229475] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 522.229786] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquiring lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.229912] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Acquired lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.230088] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 522.322251] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 522.353736] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.354039] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.354201] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.354387] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.354528] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.354941] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.354941] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.355123] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.355258] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.355421] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.355621] env[63418]: DEBUG nova.virt.hardware [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.356544] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abadeab7-d747-46e4-b5ae-76a2b1806892 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.370293] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59c4ec7-88eb-456a-ab01-ba445daddbcb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.420835] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244606, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067808} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.420835] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 522.421353] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dd62b1-e3c8-437a-baf0-0cafe8f48329 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.446818] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 522.450884] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquiring lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.451096] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquired lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.451269] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 522.453387] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a80fd7c5-6707-4e06-becb-bf88d0473b58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.476222] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 522.476222] env[63418]: value = "task-1244607" [ 522.476222] env[63418]: _type = "Task" [ 522.476222] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.492059] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244607, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.648494] env[63418]: DEBUG oslo_vmware.rw_handles [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 522.648876] env[63418]: DEBUG oslo_vmware.rw_handles [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 522.772921] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Downloaded image file data c0dccf29-5e49-4a1e-b51f-d46e566b4772 to vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk on the data store datastore1 {{(pid=63418) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 522.775154] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 522.775434] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copying Virtual Disk [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk to [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 522.775715] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c22e4950-3a6e-4690-915d-290ea8e85fd7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.779147] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.785645] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 522.785645] env[63418]: value = "task-1244608" [ 522.785645] env[63418]: _type = "Task" [ 522.785645] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.992459] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244607, 'name': ReconfigVM_Task, 'duration_secs': 0.308612} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.992760] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Reconfigured VM instance instance-00000004 to attach disk [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 522.993525] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35e6fb18-a195-4deb-9007-73c3866cd823 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.001185] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 523.001185] env[63418]: value = "task-1244609" [ 523.001185] env[63418]: _type = "Task" [ 523.001185] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.011609] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244609, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.100536] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.256025] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.300565] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244608, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.325568] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.515920] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244609, 'name': Rename_Task, 'duration_secs': 0.141407} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.515920] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 523.515920] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88adfc73-2fe5-4a4d-aa9d-822377bd6c2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.522642] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 523.522642] env[63418]: value = "task-1244610" [ 523.522642] env[63418]: _type = "Task" [ 523.522642] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.540334] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244610, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.589535] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e7b731-4a71-4a73-aab9-111ec81b54df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.595767] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8194c261-d8da-4fd5-85b8-08cb36c958b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.631401] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe46850-36a3-419b-8b9b-0c3eed550307 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.639557] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2392f778-e90d-4bb1-8105-e7c5125f58bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.655379] env[63418]: DEBUG nova.compute.provider_tree [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.759214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Releasing lock "refresh_cache-8879ea02-d856-4233-a5e8-312701705119" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.759544] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 523.759641] env[63418]: DEBUG nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 523.759816] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 523.778159] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.797880] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244608, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.728339} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.798203] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copied Virtual Disk [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk to [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 523.798332] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleting the datastore file [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772/tmp-sparse.vmdk {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 523.798579] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2572171d-e4c2-4c37-a8a7-71fd40f85113 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.805741] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 523.805741] env[63418]: value = "task-1244611" [ 523.805741] env[63418]: _type = "Task" [ 523.805741] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.815445] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.831042] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Releasing lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.831478] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 523.831741] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 523.832051] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e129556d-b2be-482e-9079-ecb552d83deb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.844023] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ae1199-099e-4c7a-b06b-7f0883d925bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.869878] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8fbc105f-34f9-4ece-9e74-e473f310221a could not be found. [ 523.869878] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 523.869878] env[63418]: INFO nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 523.869878] env[63418]: DEBUG oslo.service.loopingcall [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.869878] env[63418]: DEBUG nova.compute.manager [-] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 523.869878] env[63418]: DEBUG nova.network.neutron [-] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 523.892870] env[63418]: DEBUG nova.network.neutron [-] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.035702] env[63418]: DEBUG oslo_vmware.api [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244610, 'name': PowerOnVM_Task, 'duration_secs': 0.49421} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.036046] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 524.036520] env[63418]: INFO nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Took 8.66 seconds to spawn the instance on the hypervisor. [ 524.036520] env[63418]: DEBUG nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 524.037692] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0805d5a-dbfa-4a14-8c51-26680df886a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.158997] env[63418]: DEBUG nova.scheduler.client.report [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 524.279697] env[63418]: DEBUG nova.network.neutron [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.318855] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024522} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.318985] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 524.320199] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Moving file from [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a/c0dccf29-5e49-4a1e-b51f-d46e566b4772 to [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772. {{(pid=63418) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 524.321550] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-cc7464e3-308f-4de2-89f6-f70813ff551c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.326804] env[63418]: DEBUG nova.compute.manager [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Received event network-changed-0bccf0ff-e310-44de-a851-c3dffcfd8e8a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 524.326804] env[63418]: DEBUG nova.compute.manager [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Refreshing instance network info cache due to event network-changed-0bccf0ff-e310-44de-a851-c3dffcfd8e8a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 524.326997] env[63418]: DEBUG oslo_concurrency.lockutils [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] Acquiring lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.327275] env[63418]: DEBUG oslo_concurrency.lockutils [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] Acquired lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.327462] env[63418]: DEBUG nova.network.neutron [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Refreshing network info cache for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 524.336055] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 524.336055] env[63418]: value = "task-1244612" [ 524.336055] env[63418]: _type = "Task" [ 524.336055] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.347665] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244612, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.397056] env[63418]: DEBUG nova.network.neutron [-] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.562080] env[63418]: INFO nova.compute.manager [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Took 18.67 seconds to build instance. [ 524.667057] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.667057] env[63418]: DEBUG nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 524.668305] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.408s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.668769] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.669098] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 524.669686] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.488s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.672338] env[63418]: INFO nova.compute.claims [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.677383] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f88e73e-ab78-4900-9397-010218021156 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.689200] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856aae67-6f1f-45d3-9580-a2df568ccefa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.707803] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4cc9ae-f431-4391-9b3d-72953c1e34cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.716596] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4663a70d-c797-4416-87b4-4af91dbc7299 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.754128] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181545MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 524.754581] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.783780] env[63418]: INFO nova.compute.manager [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] [instance: 8879ea02-d856-4233-a5e8-312701705119] Took 1.02 seconds to deallocate network for instance. [ 524.848923] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244612, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025677} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.848923] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] File moved {{(pid=63418) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 524.848923] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Cleaning up location [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 524.848923] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleting the datastore file [datastore1] vmware_temp/ecb6c828-2462-4ddb-a37d-9a0760d24e0a {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 524.848923] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ac48001-c97f-4b3d-8f2b-53cd5f87c640 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.852114] env[63418]: DEBUG nova.network.neutron [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.855515] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 524.855515] env[63418]: value = "task-1244613" [ 524.855515] env[63418]: _type = "Task" [ 524.855515] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.863334] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244613, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.900199] env[63418]: INFO nova.compute.manager [-] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Took 1.03 seconds to deallocate network for instance. [ 524.903859] env[63418]: DEBUG nova.compute.claims [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 524.904089] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.960135] env[63418]: ERROR nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 524.960135] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 524.960135] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.960135] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.960135] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.960135] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.960135] env[63418]: ERROR nova.compute.manager raise self.value [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.960135] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.960135] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.960135] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.960617] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.960617] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.960617] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 524.960617] env[63418]: ERROR nova.compute.manager [ 524.960617] env[63418]: Traceback (most recent call last): [ 524.960617] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.960617] env[63418]: listener.cb(fileno) [ 524.960617] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.960617] env[63418]: result = function(*args, **kwargs) [ 524.960617] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.960617] env[63418]: return func(*args, **kwargs) [ 524.960617] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 524.960617] env[63418]: raise e [ 524.960617] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 524.960617] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 524.960617] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.960617] env[63418]: created_port_ids = self._update_ports_for_instance( [ 524.960617] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.960617] env[63418]: with excutils.save_and_reraise_exception(): [ 524.960617] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.960617] env[63418]: self.force_reraise() [ 524.960617] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.960617] env[63418]: raise self.value [ 524.960617] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.960617] env[63418]: updated_port = self._update_port( [ 524.960617] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.960617] env[63418]: _ensure_no_port_binding_failure(port) [ 524.960617] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.960617] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.961575] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 524.961575] env[63418]: Removing descriptor: 16 [ 524.961575] env[63418]: ERROR nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Traceback (most recent call last): [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] yield resources [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self.driver.spawn(context, instance, image_meta, [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.961575] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] vm_ref = self.build_virtual_machine(instance, [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] for vif in network_info: [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return self._sync_wrapper(fn, *args, **kwargs) [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self.wait() [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self[:] = self._gt.wait() [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return self._exit_event.wait() [ 524.961903] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] result = hub.switch() [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return self.greenlet.switch() [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] result = function(*args, **kwargs) [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return func(*args, **kwargs) [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] raise e [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] nwinfo = self.network_api.allocate_for_instance( [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.962312] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] created_port_ids = self._update_ports_for_instance( [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] with excutils.save_and_reraise_exception(): [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self.force_reraise() [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] raise self.value [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] updated_port = self._update_port( [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] _ensure_no_port_binding_failure(port) [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.962664] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] raise exception.PortBindingFailed(port_id=port['id']) [ 524.963019] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 524.963019] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] [ 524.963019] env[63418]: INFO nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Terminating instance [ 524.970766] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquiring lock "f714b41e-b537-4feb-bd71-53b5db089485" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.971010] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Lock "f714b41e-b537-4feb-bd71-53b5db089485" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.972546] env[63418]: DEBUG nova.network.neutron [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.071169] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4191ba22-e57e-4691-9540-426f23b57e57 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "ba67658a-668e-4fca-aefe-e838f7b05e2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.188s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.180313] env[63418]: DEBUG nova.compute.utils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.181990] env[63418]: DEBUG nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 525.368380] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244613, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.030071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.368659] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 525.369433] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82e453ac-417f-4bd5-89ad-1dcb8e61de7c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.375320] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 525.375320] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f7768c-c870-e45c-e5c3-1f82f16f9e28" [ 525.375320] env[63418]: _type = "Task" [ 525.375320] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.389886] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f7768c-c870-e45c-e5c3-1f82f16f9e28, 'name': SearchDatastore_Task, 'duration_secs': 0.008443} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.390145] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.390524] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 7e43d259-f361-43d8-8f03-72b303680478/7e43d259-f361-43d8-8f03-72b303680478.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 525.392029] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed88d395-dd00-4e2c-97f3-77a6748c93b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.399545] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 525.399545] env[63418]: value = "task-1244614" [ 525.399545] env[63418]: _type = "Task" [ 525.399545] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.408732] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244614, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.471789] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquiring lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.472012] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquired lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.472373] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 525.474767] env[63418]: DEBUG oslo_concurrency.lockutils [req-914bde3b-275a-41aa-9c3f-aa8dcfc5ba31 req-16e87341-3d2f-4703-bcf9-cc902ff3ace3 service nova] Releasing lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.573062] env[63418]: DEBUG nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 525.686725] env[63418]: DEBUG nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 525.819121] env[63418]: INFO nova.scheduler.client.report [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Deleted allocations for instance 8879ea02-d856-4233-a5e8-312701705119 [ 525.915895] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244614, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466637} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.918755] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 7e43d259-f361-43d8-8f03-72b303680478/7e43d259-f361-43d8-8f03-72b303680478.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 525.920687] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 525.923400] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c912477-c809-48e1-8391-ec24528945c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.931728] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 525.931728] env[63418]: value = "task-1244615" [ 525.931728] env[63418]: _type = "Task" [ 525.931728] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.950879] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244615, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.001801] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.022033] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5d421a-69a1-4e57-b9bc-a625979ad264 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.031667] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525ef4d6-761e-4e7d-8582-0bbedf3e9bad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.039794] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquiring lock "3905f544-be5d-4a2a-a9c6-c7421e41b2f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.039794] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Lock "3905f544-be5d-4a2a-a9c6-c7421e41b2f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.069637] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bdd416-119d-4d4c-bf85-24ec250ce744 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.077733] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecebd6c-5e32-44f5-804d-d367014e0ddc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.097442] env[63418]: DEBUG nova.compute.provider_tree [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.110445] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.141416] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.201470] env[63418]: DEBUG nova.compute.manager [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Received event network-changed-1d10ead4-73ff-4946-bf33-e3e9a105b862 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 526.201698] env[63418]: DEBUG nova.compute.manager [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Refreshing instance network info cache due to event network-changed-1d10ead4-73ff-4946-bf33-e3e9a105b862. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 526.201845] env[63418]: DEBUG oslo_concurrency.lockutils [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] Acquiring lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.339213] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3fdd02d1-24c3-4bc3-8f40-87b412f1bbae tempest-FloatingIPsAssociationNegativeTestJSON-37281641 tempest-FloatingIPsAssociationNegativeTestJSON-37281641-project-member] Lock "8879ea02-d856-4233-a5e8-312701705119" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.011s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.370562] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquiring lock "67b5708b-8d02-4816-9455-ea3d9414998a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.370835] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Lock "67b5708b-8d02-4816-9455-ea3d9414998a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.440665] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244615, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061644} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.440947] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 526.441889] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d08221-fee9-4a55-8d97-71870f142c2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.463768] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 7e43d259-f361-43d8-8f03-72b303680478/7e43d259-f361-43d8-8f03-72b303680478.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 526.463952] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f3184ef-a487-4fd2-b34f-a5ad74d1799e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.483280] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 526.483280] env[63418]: value = "task-1244616" [ 526.483280] env[63418]: _type = "Task" [ 526.483280] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.491782] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244616, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.600865] env[63418]: DEBUG nova.scheduler.client.report [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 526.645209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Releasing lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.645769] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 526.645975] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 526.646609] env[63418]: DEBUG oslo_concurrency.lockutils [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] Acquired lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.646609] env[63418]: DEBUG nova.network.neutron [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Refreshing network info cache for port 1d10ead4-73ff-4946-bf33-e3e9a105b862 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 526.651136] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-113d00b1-f999-448d-8316-312b0f591f6d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.660299] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676c2201-76ee-44fc-872f-1273026c93c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.688079] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0 could not be found. [ 526.688333] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 526.688496] env[63418]: INFO nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 526.691120] env[63418]: DEBUG oslo.service.loopingcall [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.691120] env[63418]: DEBUG nova.compute.manager [-] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 526.691120] env[63418]: DEBUG nova.network.neutron [-] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 526.704982] env[63418]: DEBUG nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 526.741796] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 526.741796] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 526.742124] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.742124] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 526.742213] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.742317] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 526.742579] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 526.742672] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 526.742841] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 526.743525] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 526.743734] env[63418]: DEBUG nova.virt.hardware [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 526.744693] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669c7303-e178-431d-a00e-d2725d2bed6f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.753554] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5288b01f-4426-4d50-8858-6180147ecf77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.769950] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 526.776703] env[63418]: DEBUG oslo.service.loopingcall [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.776994] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 526.777301] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05ee0b95-06bb-4fe6-a560-a8bff3ac40f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.795028] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 526.795028] env[63418]: value = "task-1244617" [ 526.795028] env[63418]: _type = "Task" [ 526.795028] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.804312] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244617, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.844595] env[63418]: DEBUG nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 526.879846] env[63418]: DEBUG nova.network.neutron [-] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.000732] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244616, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.109096] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.109651] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 527.116469] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.442s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.182179] env[63418]: DEBUG nova.network.neutron [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.246706] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquiring lock "58850986-3ebf-465f-b832-9b245742a34a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.246706] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Lock "58850986-3ebf-465f-b832-9b245742a34a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.285904] env[63418]: DEBUG nova.network.neutron [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.310379] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244617, 'name': CreateVM_Task, 'duration_secs': 0.322309} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.310379] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 527.310379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.310379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.310379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 527.310379] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86940273-6550-4d2b-aa93-a40eb97c9ade {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.315420] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 527.315420] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525b5088-c3c8-fa53-1c38-c0edee245a3d" [ 527.315420] env[63418]: _type = "Task" [ 527.315420] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.325869] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525b5088-c3c8-fa53-1c38-c0edee245a3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.368181] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.382467] env[63418]: DEBUG nova.network.neutron [-] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.495190] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244616, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.619741] env[63418]: DEBUG nova.compute.utils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.624365] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 527.624584] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 527.710537] env[63418]: DEBUG nova.policy [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5736c3bfbb4806b6aaef5a6e05ca7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34143fc5e0144a8fb736d3deca6cc20b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.789281] env[63418]: DEBUG oslo_concurrency.lockutils [req-0793cd4e-4c94-4594-9493-529275262b9a req-d2990116-08b5-4a32-ae42-51f4a138a73c service nova] Releasing lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.827769] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525b5088-c3c8-fa53-1c38-c0edee245a3d, 'name': SearchDatastore_Task, 'duration_secs': 0.009217} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.830416] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.830571] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 527.830905] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.831064] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.831240] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 527.831676] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-370413e6-0569-42ad-9014-7b8df9252f84 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.841878] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 527.842200] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 527.845469] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7713cc3e-cb58-4633-841d-b2f7992f0bcb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.851333] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 527.851333] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5296cac7-b9fb-c55f-d802-6884891cb1ba" [ 527.851333] env[63418]: _type = "Task" [ 527.851333] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.859155] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5296cac7-b9fb-c55f-d802-6884891cb1ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.873665] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7555791d-62e8-4a5d-99a5-16c56191ee60 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.880678] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2671b99-57f5-4b4a-b97a-1ae6b4323f58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.884624] env[63418]: INFO nova.compute.manager [-] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Took 1.20 seconds to deallocate network for instance. [ 527.911251] env[63418]: DEBUG nova.compute.claims [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 527.911436] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.912236] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c7f2dd-5aba-449d-8e95-0435ed9beb35 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.922710] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c949020-317e-40cd-b848-1ad382283d12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.937760] env[63418]: DEBUG nova.compute.provider_tree [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.998476] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244616, 'name': ReconfigVM_Task, 'duration_secs': 1.497536} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.998935] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 7e43d259-f361-43d8-8f03-72b303680478/7e43d259-f361-43d8-8f03-72b303680478.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 527.999624] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1dd03f21-f79f-4632-b5b0-8163c2c61eba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.006823] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 528.006823] env[63418]: value = "task-1244618" [ 528.006823] env[63418]: _type = "Task" [ 528.006823] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.016190] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244618, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.125277] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 528.155534] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "34cf7345-c244-44bd-aa1f-4788ae8f6da5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.155682] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "34cf7345-c244-44bd-aa1f-4788ae8f6da5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.169970] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Successfully created port: 36803e73-f0ee-4ef3-b808-d0f51174734c {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.364229] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5296cac7-b9fb-c55f-d802-6884891cb1ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008242} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.365135] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19760532-0a15-4f08-8f06-66498e09b95c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.372768] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 528.372768] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527a7b46-ecd7-9449-3d34-dc7030770a8e" [ 528.372768] env[63418]: _type = "Task" [ 528.372768] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.384534] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527a7b46-ecd7-9449-3d34-dc7030770a8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.441430] env[63418]: DEBUG nova.scheduler.client.report [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 528.517854] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244618, 'name': Rename_Task, 'duration_secs': 0.249568} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.520862] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 528.521165] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d3db448-c2c0-44c7-816f-bda55cba8ed5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.532367] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 528.532367] env[63418]: value = "task-1244619" [ 528.532367] env[63418]: _type = "Task" [ 528.532367] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.538774] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244619, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.689830] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Successfully created port: 7c19c716-8637-4438-8917-68682f7297ff {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.883544] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527a7b46-ecd7-9449-3d34-dc7030770a8e, 'name': SearchDatastore_Task, 'duration_secs': 0.01033} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.883823] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.884102] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 528.884356] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90053f75-6a6e-4863-b575-c88bd6cb1dd2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.890639] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 528.890639] env[63418]: value = "task-1244620" [ 528.890639] env[63418]: _type = "Task" [ 528.890639] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.899257] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.947920] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.834s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.949015] env[63418]: ERROR nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Traceback (most recent call last): [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self.driver.spawn(context, instance, image_meta, [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] vm_ref = self.build_virtual_machine(instance, [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.949015] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] for vif in network_info: [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return self._sync_wrapper(fn, *args, **kwargs) [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self.wait() [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self[:] = self._gt.wait() [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return self._exit_event.wait() [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] result = hub.switch() [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.949454] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return self.greenlet.switch() [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] result = function(*args, **kwargs) [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] return func(*args, **kwargs) [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] raise e [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] nwinfo = self.network_api.allocate_for_instance( [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] created_port_ids = self._update_ports_for_instance( [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] with excutils.save_and_reraise_exception(): [ 528.949813] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] self.force_reraise() [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] raise self.value [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] updated_port = self._update_port( [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] _ensure_no_port_binding_failure(port) [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] raise exception.PortBindingFailed(port_id=port['id']) [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] nova.exception.PortBindingFailed: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. [ 528.950153] env[63418]: ERROR nova.compute.manager [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] [ 528.950504] env[63418]: DEBUG nova.compute.utils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 528.950544] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.956758] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Build of instance 9122ae95-8678-48bd-9299-b217c77e7a0d was re-scheduled: Binding failed for port acde18ca-d2f2-4f91-ab71-cf0bdec52b55, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 528.956758] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 528.956758] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquiring lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.956758] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Acquired lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.956980] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.043468] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.135849] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 529.167050] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.167351] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.167507] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.167682] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.167821] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.168017] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.168474] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.168541] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.168708] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.168868] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.169105] env[63418]: DEBUG nova.virt.hardware [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.170243] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c1644b-5e78-4bd8-9180-e039de243940 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.180992] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b01843d-85e2-4685-a7e1-cc1415947f35 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.325262] env[63418]: DEBUG nova.compute.manager [req-309387ef-88e6-4d1e-84a5-3a0c7d55e038 req-2fe76b2a-0e14-4cbd-952c-815d751cc13f service nova] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Received event network-vif-deleted-0bccf0ff-e310-44de-a851-c3dffcfd8e8a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 529.335040] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Successfully created port: 333cf339-ae87-4872-9e9d-f17d6d893d02 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.401309] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244620, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454448} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.401591] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 529.401793] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 529.402051] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92ad1006-c030-4e4d-bc00-25b26b205676 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.409343] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 529.409343] env[63418]: value = "task-1244621" [ 529.409343] env[63418]: _type = "Task" [ 529.409343] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.418437] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244621, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.491428] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.543170] env[63418]: DEBUG oslo_vmware.api [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244619, 'name': PowerOnVM_Task, 'duration_secs': 0.714091} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.543290] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 529.543504] env[63418]: INFO nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Took 9.55 seconds to spawn the instance on the hypervisor. [ 529.544337] env[63418]: DEBUG nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 529.547765] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eeb0101-e5ff-4aec-8a99-6c8e89a076de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.613116] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.756225] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e6d9c6-a413-43fd-87a5-f4896300f985 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.764477] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc089fe4-09aa-48d9-94e5-b441c4a9bab9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.806527] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9c4045-3642-4ef8-a03b-b2f7676f08ef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.815584] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc6d4c9-f274-40a2-a4f1-e4186501f5db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.833820] env[63418]: DEBUG nova.compute.provider_tree [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.920215] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244621, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066008} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.920476] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 529.921599] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fbd9ba-4387-46d8-ac51-21cb7c23ec64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.943715] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 529.944145] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49b3baf3-b9dd-4ce2-ab3c-293a426e8570 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.970256] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 529.970256] env[63418]: value = "task-1244622" [ 529.970256] env[63418]: _type = "Task" [ 529.970256] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.978977] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244622, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.069933] env[63418]: INFO nova.compute.manager [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Took 16.41 seconds to build instance. [ 530.115061] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Releasing lock "refresh_cache-9122ae95-8678-48bd-9299-b217c77e7a0d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.118019] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 530.118019] env[63418]: DEBUG nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 530.118019] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 530.170440] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.336981] env[63418]: DEBUG nova.scheduler.client.report [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 530.483999] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244622, 'name': ReconfigVM_Task, 'duration_secs': 0.279051} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.484559] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Reconfigured VM instance instance-00000008 to attach disk [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 530.484947] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a689b3d1-11d4-475f-a499-85083334c0be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.494504] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 530.494504] env[63418]: value = "task-1244623" [ 530.494504] env[63418]: _type = "Task" [ 530.494504] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.503848] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244623, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.572260] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a540af94-ed8e-46b8-bfce-765d004d2de9 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "7e43d259-f361-43d8-8f03-72b303680478" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.927s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.672734] env[63418]: DEBUG nova.network.neutron [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.843399] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.843399] env[63418]: ERROR nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Traceback (most recent call last): [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self.driver.spawn(context, instance, image_meta, [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 530.843399] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] vm_ref = self.build_virtual_machine(instance, [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] for vif in network_info: [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return self._sync_wrapper(fn, *args, **kwargs) [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self.wait() [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self[:] = self._gt.wait() [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return self._exit_event.wait() [ 530.843731] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] result = hub.switch() [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return self.greenlet.switch() [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] result = function(*args, **kwargs) [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] return func(*args, **kwargs) [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] raise e [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] nwinfo = self.network_api.allocate_for_instance( [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.844082] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] created_port_ids = self._update_ports_for_instance( [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] with excutils.save_and_reraise_exception(): [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] self.force_reraise() [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] raise self.value [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] updated_port = self._update_port( [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] _ensure_no_port_binding_failure(port) [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.844403] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] raise exception.PortBindingFailed(port_id=port['id']) [ 530.844689] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] nova.exception.PortBindingFailed: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. [ 530.844689] env[63418]: ERROR nova.compute.manager [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] [ 530.844689] env[63418]: DEBUG nova.compute.utils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 530.848104] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Build of instance 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf was re-scheduled: Binding failed for port 76f691d8-a4d1-409c-9ff2-b21e7dea3ac1, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 530.848104] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 530.848104] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.848104] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquired lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.848859] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 530.851580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.610s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.853519] env[63418]: INFO nova.compute.claims [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.006971] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244623, 'name': Rename_Task, 'duration_secs': 0.129152} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.008895] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 531.009171] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a289dbb-789d-4bcc-8299-cce626fe783f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.013872] env[63418]: DEBUG nova.compute.manager [req-eb80926b-577e-427f-b708-6cdd7d608240 req-4f403941-0f62-401e-8d18-7fcfb6e7d992 service nova] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Received event network-vif-deleted-1d10ead4-73ff-4946-bf33-e3e9a105b862 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 531.018563] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 531.018563] env[63418]: value = "task-1244624" [ 531.018563] env[63418]: _type = "Task" [ 531.018563] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.031020] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.075753] env[63418]: DEBUG nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 531.177019] env[63418]: INFO nova.compute.manager [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] [instance: 9122ae95-8678-48bd-9299-b217c77e7a0d] Took 1.06 seconds to deallocate network for instance. [ 531.380165] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.457570] env[63418]: INFO nova.compute.manager [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Rebuilding instance [ 531.476090] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.511233] env[63418]: DEBUG nova.compute.manager [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 531.512371] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1c2be8-0fb5-4153-82c2-472309d1ca89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.535233] env[63418]: DEBUG oslo_vmware.api [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244624, 'name': PowerOnVM_Task, 'duration_secs': 0.449024} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.536759] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 531.536759] env[63418]: INFO nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Took 4.83 seconds to spawn the instance on the hypervisor. [ 531.536974] env[63418]: DEBUG nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 531.537852] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab27d4a5-8303-4357-a602-6c469f20bbdc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.560143] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquiring lock "5caab97d-833e-4356-af51-4a991d0e5b44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.560143] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Lock "5caab97d-833e-4356-af51-4a991d0e5b44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.606502] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.756202] env[63418]: ERROR nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 531.756202] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 531.756202] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 531.756202] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 531.756202] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.756202] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.756202] env[63418]: ERROR nova.compute.manager raise self.value [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 531.756202] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 531.756202] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.756202] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 531.757813] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.757813] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 531.757813] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 531.757813] env[63418]: ERROR nova.compute.manager [ 531.757813] env[63418]: Traceback (most recent call last): [ 531.757813] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 531.757813] env[63418]: listener.cb(fileno) [ 531.757813] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.757813] env[63418]: result = function(*args, **kwargs) [ 531.757813] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.757813] env[63418]: return func(*args, **kwargs) [ 531.757813] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 531.757813] env[63418]: raise e [ 531.757813] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 531.757813] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 531.757813] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 531.757813] env[63418]: created_port_ids = self._update_ports_for_instance( [ 531.757813] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 531.757813] env[63418]: with excutils.save_and_reraise_exception(): [ 531.757813] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.757813] env[63418]: self.force_reraise() [ 531.757813] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.757813] env[63418]: raise self.value [ 531.757813] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 531.757813] env[63418]: updated_port = self._update_port( [ 531.757813] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.757813] env[63418]: _ensure_no_port_binding_failure(port) [ 531.757813] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.757813] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 531.758602] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 531.758602] env[63418]: Removing descriptor: 15 [ 531.758602] env[63418]: ERROR nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Traceback (most recent call last): [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] yield resources [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self.driver.spawn(context, instance, image_meta, [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 531.758602] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] vm_ref = self.build_virtual_machine(instance, [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] vif_infos = vmwarevif.get_vif_info(self._session, [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] for vif in network_info: [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return self._sync_wrapper(fn, *args, **kwargs) [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self.wait() [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self[:] = self._gt.wait() [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return self._exit_event.wait() [ 531.759015] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] result = hub.switch() [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return self.greenlet.switch() [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] result = function(*args, **kwargs) [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return func(*args, **kwargs) [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] raise e [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] nwinfo = self.network_api.allocate_for_instance( [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 531.759411] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] created_port_ids = self._update_ports_for_instance( [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] with excutils.save_and_reraise_exception(): [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self.force_reraise() [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] raise self.value [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] updated_port = self._update_port( [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] _ensure_no_port_binding_failure(port) [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.759777] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] raise exception.PortBindingFailed(port_id=port['id']) [ 531.760510] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 531.760510] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] [ 531.760510] env[63418]: INFO nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Terminating instance [ 531.983531] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Releasing lock "refresh_cache-37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.983531] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 531.983712] env[63418]: DEBUG nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 531.985341] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.005697] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.059842] env[63418]: INFO nova.compute.manager [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Took 15.41 seconds to build instance. [ 532.186943] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d167c29f-401a-4687-a6db-e35cab202d26 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.194844] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbb0ee1-7c34-4970-8b5e-e3d42e7fafc9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.246279] env[63418]: INFO nova.scheduler.client.report [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Deleted allocations for instance 9122ae95-8678-48bd-9299-b217c77e7a0d [ 532.256518] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b42f2e-6449-4507-a87d-bc4df4c63fd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.266152] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4be2500-6ad2-439b-9c7b-4276b456de40 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.271045] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.271045] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquired lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.271156] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 532.282436] env[63418]: DEBUG nova.compute.provider_tree [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.516691] env[63418]: DEBUG nova.network.neutron [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.534797] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 532.535111] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e41e0d2d-9ea3-4a12-8725-4495bf15463d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.544036] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 532.544036] env[63418]: value = "task-1244625" [ 532.544036] env[63418]: _type = "Task" [ 532.544036] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.553459] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.563488] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c62bf269-fe66-4865-9035-630b489ecb91 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "f185a348-e91d-48d1-970b-473cc253cfdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.921s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.742566] env[63418]: DEBUG nova.compute.manager [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Received event network-changed-36803e73-f0ee-4ef3-b808-d0f51174734c {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 532.742566] env[63418]: DEBUG nova.compute.manager [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Refreshing instance network info cache due to event network-changed-36803e73-f0ee-4ef3-b808-d0f51174734c. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 532.742566] env[63418]: DEBUG oslo_concurrency.lockutils [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] Acquiring lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.761464] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0738302b-fbd3-4416-b2f2-d9d0ac89694f tempest-ServersAdminNegativeTestJSON-1363015799 tempest-ServersAdminNegativeTestJSON-1363015799-project-member] Lock "9122ae95-8678-48bd-9299-b217c77e7a0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.505s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.785826] env[63418]: DEBUG nova.scheduler.client.report [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 532.830075] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.020192] env[63418]: INFO nova.compute.manager [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf] Took 1.04 seconds to deallocate network for instance. [ 533.055385] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244625, 'name': PowerOffVM_Task, 'duration_secs': 0.416802} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.055670] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 533.055910] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 533.056812] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25bb3dc4-c9b0-49e0-abf7-4cb17587dceb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.064228] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 533.064483] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-600dd399-d327-485d-8e9d-d6f6db5603fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.068205] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 533.096217] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 533.096468] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 533.096697] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleting the datastore file [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 533.096894] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ba85a4b-097a-43e2-b84e-f14ddb1dfc4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.104180] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 533.104180] env[63418]: value = "task-1244627" [ 533.104180] env[63418]: _type = "Task" [ 533.104180] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.116018] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244627, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.130916] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.264386] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 533.292612] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.293217] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 533.298798] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.544s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.395459] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquiring lock "b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.395699] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Lock "b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.590384] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.614273] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244627, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199488} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.614516] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 533.614696] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 533.614941] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.634541] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Releasing lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.635235] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 533.635235] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 533.637967] env[63418]: DEBUG oslo_concurrency.lockutils [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] Acquired lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.637967] env[63418]: DEBUG nova.network.neutron [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Refreshing network info cache for port 36803e73-f0ee-4ef3-b808-d0f51174734c {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 533.637967] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e50b468e-1227-49bd-a27d-1938baa26915 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.652246] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29763654-3f9f-41df-a56b-5d820fe50c68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.680297] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d255e7af-b50d-4b30-8bec-9e2ee2de2c80 could not be found. [ 533.680575] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.680761] env[63418]: INFO nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Took 0.05 seconds to destroy the instance on the hypervisor. [ 533.680993] env[63418]: DEBUG oslo.service.loopingcall [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.681643] env[63418]: DEBUG nova.compute.manager [-] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 533.681738] env[63418]: DEBUG nova.network.neutron [-] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 533.765262] env[63418]: DEBUG nova.network.neutron [-] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.785598] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.801728] env[63418]: DEBUG nova.compute.utils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.803814] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 533.804183] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 533.843530] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquiring lock "a0584abb-a226-4f53-b01f-faee7ffcbf48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.843762] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Lock "a0584abb-a226-4f53-b01f-faee7ffcbf48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.962782] env[63418]: DEBUG nova.policy [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e14c925f44747679e4d1299f089ceca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2e38c6b054f47ea88f35f748423b7d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.061666] env[63418]: INFO nova.scheduler.client.report [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Deleted allocations for instance 37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf [ 534.216110] env[63418]: DEBUG nova.network.neutron [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.308755] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 534.343628] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance ba67658a-668e-4fca-aefe-e838f7b05e2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.343723] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 8fbc105f-34f9-4ece-9e74-e473f310221a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.343880] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7e43d259-f361-43d8-8f03-72b303680478 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.343955] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.344466] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance f185a348-e91d-48d1-970b-473cc253cfdf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.344466] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance d255e7af-b50d-4b30-8bec-9e2ee2de2c80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.344466] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 781262a7-56ba-466c-86b3-1551905c08bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 534.438859] env[63418]: DEBUG nova.network.neutron [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.486214] env[63418]: INFO nova.compute.manager [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Rebuilding instance [ 534.545032] env[63418]: DEBUG nova.compute.manager [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 534.545922] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d210eda8-ff5d-424d-ae96-0439fb9103af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.571952] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8dbddbeb-15e3-4ac1-9eeb-060f23033f43 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "37f3d5fc-c0ee-4c10-9cbc-8cac9dbd4bdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.834s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.657012] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.657267] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.657498] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.657711] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.658171] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.658365] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.658707] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.658876] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.659998] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.659998] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.659998] env[63418]: DEBUG nova.virt.hardware [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.660565] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004bd198-8314-42a9-b90f-118cf7a8d5ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.669264] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d648ef6-35d5-4078-ad5e-967bfe4c6712 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.684546] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 534.691567] env[63418]: DEBUG oslo.service.loopingcall [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.691567] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 534.691567] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39b2e908-a04c-44be-beae-5e1b013a3e49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.714408] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 534.714408] env[63418]: value = "task-1244628" [ 534.714408] env[63418]: _type = "Task" [ 534.714408] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.722569] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244628, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.762749] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Successfully created port: 902a657c-5430-4645-ab19-a5ec12fe99cf {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.848909] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 01deeb21-0e27-497f-9b85-c85949a3533d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.942618] env[63418]: DEBUG oslo_concurrency.lockutils [req-bf0358ce-f578-460b-8a1c-f1a75801b258 req-2d086690-6b5d-45ca-b18c-3f9f80a90998 service nova] Releasing lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.019207] env[63418]: DEBUG nova.network.neutron [-] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.074833] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 535.224797] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244628, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.317954] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 535.354458] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance ed637299-7e69-43d7-85f9-8fcbd6e90dec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.358488] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 535.358768] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 535.358859] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 535.359042] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 535.359232] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 535.359346] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 535.359568] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 535.359733] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 535.359907] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 535.360091] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 535.360366] env[63418]: DEBUG nova.virt.hardware [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 535.361472] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04217df3-59b6-48e9-a063-e80bb54cc8f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.371099] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af74b54f-8d2d-4dce-b1df-8be4d1ebd521 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.523381] env[63418]: INFO nova.compute.manager [-] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Took 1.84 seconds to deallocate network for instance. [ 535.524869] env[63418]: DEBUG nova.compute.claims [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.525212] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.562765] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 535.562765] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7976703f-95eb-4a3f-829b-9a4ad755a99f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.568377] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 535.568377] env[63418]: value = "task-1244629" [ 535.568377] env[63418]: _type = "Task" [ 535.568377] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.579782] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244629, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.607456] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.725942] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244628, 'name': CreateVM_Task, 'duration_secs': 0.607395} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.726362] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 535.726951] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.729225] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.729225] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 535.729225] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24062e19-8850-41f0-8553-f66ae0fded57 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.737302] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 535.737302] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b3a93e-48ef-8284-781e-1d00f06e97ea" [ 535.737302] env[63418]: _type = "Task" [ 535.737302] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.745647] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b3a93e-48ef-8284-781e-1d00f06e97ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008791} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.745986] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.746291] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 535.746558] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.746731] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.747193] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 535.747193] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0acdc4d-a36c-4fb1-9c7c-59e4dfa41d61 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.755695] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 535.755881] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 535.756644] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0292041b-de28-492a-91ca-19c1c267d2b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.762386] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 535.762386] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5233de62-542a-e383-5450-89f666a404a4" [ 535.762386] env[63418]: _type = "Task" [ 535.762386] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.772730] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5233de62-542a-e383-5450-89f666a404a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.864047] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance a8c7f192-4672-43cc-8c38-0c33ce633765 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.895641] env[63418]: DEBUG nova.compute.manager [req-ec48eb5b-0a1f-4f1d-ae94-679be4203d00 req-923569b8-b267-438b-b9d2-f61d21143db6 service nova] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Received event network-vif-deleted-36803e73-f0ee-4ef3-b808-d0f51174734c {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 536.081836] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244629, 'name': PowerOffVM_Task, 'duration_secs': 0.137918} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.082795] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 536.083141] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 536.083898] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dc476b-8c2c-4975-8f75-dfb884f52035 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.092535] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 536.092800] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be09dbc2-edef-44fe-8b90-1a17bcc5319c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.117414] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 536.117861] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 536.118087] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleting the datastore file [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 536.118355] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4aa83798-443f-4196-b065-1f3b26550032 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.125223] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 536.125223] env[63418]: value = "task-1244631" [ 536.125223] env[63418]: _type = "Task" [ 536.125223] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.136488] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.273809] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5233de62-542a-e383-5450-89f666a404a4, 'name': SearchDatastore_Task, 'duration_secs': 0.010643} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.274653] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a345b949-3c84-47a7-bb67-5d796452b222 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.280424] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 536.280424] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522d64b8-f6a2-ae60-cb76-b3eae8c5233e" [ 536.280424] env[63418]: _type = "Task" [ 536.280424] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.294309] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522d64b8-f6a2-ae60-cb76-b3eae8c5233e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.366970] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance f714b41e-b537-4feb-bd71-53b5db089485 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 536.435431] env[63418]: ERROR nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 536.435431] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.435431] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.435431] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.435431] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.435431] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.435431] env[63418]: ERROR nova.compute.manager raise self.value [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.435431] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.435431] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.435431] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.435951] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.435951] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.435951] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 536.435951] env[63418]: ERROR nova.compute.manager [ 536.435951] env[63418]: Traceback (most recent call last): [ 536.435951] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.435951] env[63418]: listener.cb(fileno) [ 536.435951] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.435951] env[63418]: result = function(*args, **kwargs) [ 536.435951] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.435951] env[63418]: return func(*args, **kwargs) [ 536.435951] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 536.435951] env[63418]: raise e [ 536.435951] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.435951] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 536.435951] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.435951] env[63418]: created_port_ids = self._update_ports_for_instance( [ 536.435951] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.435951] env[63418]: with excutils.save_and_reraise_exception(): [ 536.435951] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.435951] env[63418]: self.force_reraise() [ 536.435951] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.435951] env[63418]: raise self.value [ 536.435951] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.435951] env[63418]: updated_port = self._update_port( [ 536.435951] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.435951] env[63418]: _ensure_no_port_binding_failure(port) [ 536.435951] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.435951] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.436658] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 536.436658] env[63418]: Removing descriptor: 16 [ 536.436658] env[63418]: ERROR nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Traceback (most recent call last): [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] yield resources [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self.driver.spawn(context, instance, image_meta, [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.436658] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] vm_ref = self.build_virtual_machine(instance, [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] for vif in network_info: [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return self._sync_wrapper(fn, *args, **kwargs) [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self.wait() [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self[:] = self._gt.wait() [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return self._exit_event.wait() [ 536.436964] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] result = hub.switch() [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return self.greenlet.switch() [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] result = function(*args, **kwargs) [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return func(*args, **kwargs) [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] raise e [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] nwinfo = self.network_api.allocate_for_instance( [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.437335] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] created_port_ids = self._update_ports_for_instance( [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] with excutils.save_and_reraise_exception(): [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self.force_reraise() [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] raise self.value [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] updated_port = self._update_port( [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] _ensure_no_port_binding_failure(port) [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.437666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] raise exception.PortBindingFailed(port_id=port['id']) [ 536.440346] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 536.440346] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] [ 536.440346] env[63418]: INFO nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Terminating instance [ 536.636299] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244631, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100931} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.637769] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 536.637846] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 536.638153] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.657893] env[63418]: DEBUG nova.compute.manager [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Received event network-changed-902a657c-5430-4645-ab19-a5ec12fe99cf {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 536.658120] env[63418]: DEBUG nova.compute.manager [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Refreshing instance network info cache due to event network-changed-902a657c-5430-4645-ab19-a5ec12fe99cf. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 536.658391] env[63418]: DEBUG oslo_concurrency.lockutils [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] Acquiring lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.658481] env[63418]: DEBUG oslo_concurrency.lockutils [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] Acquired lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.658692] env[63418]: DEBUG nova.network.neutron [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Refreshing network info cache for port 902a657c-5430-4645-ab19-a5ec12fe99cf {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 536.794569] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522d64b8-f6a2-ae60-cb76-b3eae8c5233e, 'name': SearchDatastore_Task, 'duration_secs': 0.008767} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.795754] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.796183] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 536.796574] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f018249-9ff8-473b-afb4-c6841178c555 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.804338] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 536.804338] env[63418]: value = "task-1244636" [ 536.804338] env[63418]: _type = "Task" [ 536.804338] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.815511] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244636, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.872296] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 3905f544-be5d-4a2a-a9c6-c7421e41b2f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 536.942388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.314516] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244636, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473227} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.314516] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 537.314687] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 537.316145] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5a79542-aac7-4d69-b99a-29875507179f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.325029] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 537.325029] env[63418]: value = "task-1244637" [ 537.325029] env[63418]: _type = "Task" [ 537.325029] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.330264] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244637, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.347812] env[63418]: DEBUG nova.network.neutron [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.372852] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 67b5708b-8d02-4816-9455-ea3d9414998a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.435141] env[63418]: DEBUG nova.network.neutron [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.676929] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.677395] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.677706] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.678011] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.678863] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.678863] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.678863] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.678863] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.679113] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.679113] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.679272] env[63418]: DEBUG nova.virt.hardware [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.680111] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06962f1a-ccbd-45cf-be20-729b63bf6042 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.688486] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad21654-db88-4386-97b7-e6be66c9831a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.701817] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 537.707709] env[63418]: DEBUG oslo.service.loopingcall [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.708248] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 537.708466] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4070c0c-abfe-4286-b592-4dddc7299602 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.724709] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 537.724709] env[63418]: value = "task-1244638" [ 537.724709] env[63418]: _type = "Task" [ 537.724709] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.734348] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244638, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.834706] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244637, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098353} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.835199] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 537.836397] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d665d593-dd81-49dc-92c9-ddc1902250b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.859787] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 537.859907] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b38201f-f410-4710-aa5f-1fbe09147ebe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.876049] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 58850986-3ebf-465f-b832-9b245742a34a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.882022] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 537.882022] env[63418]: value = "task-1244639" [ 537.882022] env[63418]: _type = "Task" [ 537.882022] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.888688] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244639, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.940874] env[63418]: DEBUG oslo_concurrency.lockutils [req-3dcd94ca-be6d-4450-a6ed-5dd3697dd88b req-f0937726-f0e4-4659-a539-618f88883afa service nova] Releasing lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.940874] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquired lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.940874] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.239526] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244638, 'name': CreateVM_Task, 'duration_secs': 0.299648} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.239526] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 538.239526] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.239526] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.239679] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 538.239961] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2d4118d-dc64-4617-99ef-eeee6c99d41e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.245652] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 538.245652] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5246ab78-2219-48b7-6ba3-49da62380e08" [ 538.245652] env[63418]: _type = "Task" [ 538.245652] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.253601] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5246ab78-2219-48b7-6ba3-49da62380e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.383023] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 34cf7345-c244-44bd-aa1f-4788ae8f6da5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.400599] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244639, 'name': ReconfigVM_Task, 'duration_secs': 0.316794} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.400599] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Reconfigured VM instance instance-00000004 to attach disk [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 538.401744] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46624d19-e675-4ada-aebd-3e5b4e8ebf7d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.409380] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 538.409380] env[63418]: value = "task-1244640" [ 538.409380] env[63418]: _type = "Task" [ 538.409380] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.417989] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244640, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.466055] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.608941] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.757330] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5246ab78-2219-48b7-6ba3-49da62380e08, 'name': SearchDatastore_Task, 'duration_secs': 0.008047} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.758379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.758812] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 538.759200] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.759454] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.759733] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 538.760106] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca872192-bcd9-42c7-9ac3-6d07b5c679c2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.772271] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 538.772271] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 538.772271] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a068d62b-8440-4a5d-acc5-925d100c13e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.780407] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 538.780407] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525a1762-aae6-a118-377a-1d5610dc0c6a" [ 538.780407] env[63418]: _type = "Task" [ 538.780407] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.792874] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525a1762-aae6-a118-377a-1d5610dc0c6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.893665] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 5caab97d-833e-4356-af51-4a991d0e5b44 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.931035] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244640, 'name': Rename_Task, 'duration_secs': 0.126759} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.931035] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 538.931035] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1e9755b-19b9-45c0-b7f0-3ccc71c064e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.939390] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 538.939390] env[63418]: value = "task-1244641" [ 538.939390] env[63418]: _type = "Task" [ 538.939390] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.951890] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.122452] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Releasing lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.122452] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 539.122452] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.122452] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f65f3656-0471-4411-88ac-a2cfe50eeef6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.132960] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea306fd-f4d5-4259-8690-cdac26728a9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.162833] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 781262a7-56ba-466c-86b3-1551905c08bf could not be found. [ 539.162917] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.163302] env[63418]: INFO nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 539.163477] env[63418]: DEBUG oslo.service.loopingcall [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.163746] env[63418]: DEBUG nova.compute.manager [-] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 539.163805] env[63418]: DEBUG nova.network.neutron [-] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.194098] env[63418]: DEBUG nova.network.neutron [-] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.290552] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525a1762-aae6-a118-377a-1d5610dc0c6a, 'name': SearchDatastore_Task, 'duration_secs': 0.016135} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.295024] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13760d42-4c4a-47d6-93d3-02032dcccd56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.301966] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 539.301966] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52da7ac0-4236-5cb2-451b-7cda23849ec8" [ 539.301966] env[63418]: _type = "Task" [ 539.301966] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.312070] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52da7ac0-4236-5cb2-451b-7cda23849ec8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.385853] env[63418]: DEBUG nova.compute.manager [req-cec61f1a-5f14-46a4-b127-43a1211278c1 req-e01fd4f1-5810-4717-97ab-28c809266f88 service nova] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Received event network-vif-deleted-902a657c-5430-4645-ab19-a5ec12fe99cf {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 539.399901] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance b1b24de1-3d7f-40b7-9183-c4b1f35f8c94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.455879] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244641, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.595574] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "37bf3e05-e797-4437-a959-96a92d4ad4e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.596910] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "37bf3e05-e797-4437-a959-96a92d4ad4e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.695830] env[63418]: DEBUG nova.network.neutron [-] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.817032] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52da7ac0-4236-5cb2-451b-7cda23849ec8, 'name': SearchDatastore_Task, 'duration_secs': 0.037729} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.817032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.817032] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 539.817032] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79dbae68-d34a-4aa4-bba9-456286d5dff9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.823062] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 539.823062] env[63418]: value = "task-1244643" [ 539.823062] env[63418]: _type = "Task" [ 539.823062] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.833529] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244643, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.905250] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance a0584abb-a226-4f53-b01f-faee7ffcbf48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.905250] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 539.905250] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 539.956868] env[63418]: DEBUG oslo_vmware.api [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244641, 'name': PowerOnVM_Task, 'duration_secs': 0.845945} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.959715] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 539.960103] env[63418]: DEBUG nova.compute.manager [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 539.964293] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4237c0e0-f3f0-4523-a830-7f260e578069 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.198602] env[63418]: INFO nova.compute.manager [-] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Took 1.03 seconds to deallocate network for instance. [ 540.203512] env[63418]: DEBUG nova.compute.claims [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.203706] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.259475] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a912a8-127d-427d-86ec-1a6fa0d18ba9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.269767] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ee1822-987c-42c2-a2d2-28e3cf2da2d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.316818] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6434e6d0-41bc-475a-9993-ef808f9cd927 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.334940] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3337593c-8ab5-43da-86fe-a32b42813010 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.357484] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244643, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.358614] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.484551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.838755] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244643, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.864024] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 541.348024] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244643, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.046622} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.348024] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 541.348024] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 541.348024] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23a58917-c20c-4000-9479-b2834a088cc2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.356270] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 541.356270] env[63418]: value = "task-1244644" [ 541.356270] env[63418]: _type = "Task" [ 541.356270] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.370562] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 541.370718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.072s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.371157] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244644, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.371288] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.467s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.519686] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquiring lock "9b2d0dfa-6820-451f-b86d-590f89928278" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.520023] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Lock "9b2d0dfa-6820-451f-b86d-590f89928278" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.878213] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244644, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068497} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.878484] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 541.881147] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138e60bd-45a5-4a58-9104-dbf6d8f18140 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.922474] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 541.922823] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2c3d1f2-b7dc-4424-a5b9-67eb6a309c9b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.959521] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 541.959521] env[63418]: value = "task-1244646" [ 541.959521] env[63418]: _type = "Task" [ 541.959521] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.969408] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244646, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.021273] env[63418]: INFO nova.compute.manager [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Rebuilding instance [ 542.101956] env[63418]: DEBUG nova.compute.manager [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 542.102896] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e8bc68-78c4-4420-8819-41024e8286f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.338610] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9542b3d8-0759-4e0f-bd2e-5b0e8428345f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.348202] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c4f54a-dcca-4002-a263-b1aac37d391e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.387563] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f11bc7a-5fd9-4531-9ac8-7bdffe058933 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.396763] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75af494-bca9-452e-9e72-e0d4b1749631 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.416097] env[63418]: DEBUG nova.compute.provider_tree [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.470919] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244646, 'name': ReconfigVM_Task, 'duration_secs': 0.471131} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.471169] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Reconfigured VM instance instance-00000008 to attach disk [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf/f185a348-e91d-48d1-970b-473cc253cfdf.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 542.471832] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fffd80a2-4385-4baa-8403-c05b7bfcc5d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.481117] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 542.481117] env[63418]: value = "task-1244648" [ 542.481117] env[63418]: _type = "Task" [ 542.481117] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.488787] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244648, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.922532] env[63418]: DEBUG nova.scheduler.client.report [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 542.991612] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244648, 'name': Rename_Task, 'duration_secs': 0.250467} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.992131] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 542.992459] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3aa59eaa-5550-4df4-8494-ba9cd083eb23 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.003185] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 543.003185] env[63418]: value = "task-1244649" [ 543.003185] env[63418]: _type = "Task" [ 543.003185] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.010019] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.120197] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 543.120720] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f424163b-012b-433e-b536-290f59f64efa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.129802] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 543.129802] env[63418]: value = "task-1244650" [ 543.129802] env[63418]: _type = "Task" [ 543.129802] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.139902] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.429618] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.058s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.430963] env[63418]: ERROR nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Traceback (most recent call last): [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self.driver.spawn(context, instance, image_meta, [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] vm_ref = self.build_virtual_machine(instance, [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.430963] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] for vif in network_info: [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return self._sync_wrapper(fn, *args, **kwargs) [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self.wait() [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self[:] = self._gt.wait() [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return self._exit_event.wait() [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] result = hub.switch() [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.431338] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return self.greenlet.switch() [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] result = function(*args, **kwargs) [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] return func(*args, **kwargs) [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] raise e [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] nwinfo = self.network_api.allocate_for_instance( [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] created_port_ids = self._update_ports_for_instance( [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] with excutils.save_and_reraise_exception(): [ 543.431711] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] self.force_reraise() [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] raise self.value [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] updated_port = self._update_port( [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] _ensure_no_port_binding_failure(port) [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] raise exception.PortBindingFailed(port_id=port['id']) [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] nova.exception.PortBindingFailed: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. [ 543.432129] env[63418]: ERROR nova.compute.manager [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] [ 543.432462] env[63418]: DEBUG nova.compute.utils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.433965] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.323s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.435363] env[63418]: INFO nova.compute.claims [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.442555] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Build of instance 8fbc105f-34f9-4ece-9e74-e473f310221a was re-scheduled: Binding failed for port 0bccf0ff-e310-44de-a851-c3dffcfd8e8a, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 543.443069] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 543.443358] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquiring lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.443507] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Acquired lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.443659] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.517970] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244649, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.640244] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244650, 'name': PowerOffVM_Task, 'duration_secs': 0.121421} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.641567] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 543.641567] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 543.641680] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adea147e-8dd8-4527-9ac7-6b9029cd211d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.652869] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 543.652869] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9d2a57d-321f-4205-bfb9-97c49272c54a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.679056] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 543.679391] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 543.679781] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Deleting the datastore file [datastore2] ba67658a-668e-4fca-aefe-e838f7b05e2a {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 543.683201] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0126e5c-c6d9-42b4-9d55-59ee0a558d4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.689141] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 543.689141] env[63418]: value = "task-1244652" [ 543.689141] env[63418]: _type = "Task" [ 543.689141] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.698105] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244652, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.976268] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.012488] env[63418]: DEBUG oslo_vmware.api [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244649, 'name': PowerOnVM_Task, 'duration_secs': 0.550936} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.012931] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 544.013280] env[63418]: DEBUG nova.compute.manager [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 544.014276] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30094ca2-2253-4bac-893f-66329056d322 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.179717] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.203349] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244652, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243015} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.203810] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 544.204138] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 544.204346] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.544724] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.680193] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Releasing lock "refresh_cache-8fbc105f-34f9-4ece-9e74-e473f310221a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.680193] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 544.680193] env[63418]: DEBUG nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 544.680193] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.764483] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.943077] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2bdb3c-1490-47c0-a827-f831272128ff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.953165] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9905dbf-7db6-4871-a076-fcae6212a6ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.997292] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f20bced-516a-400d-b37b-18f9890ac0d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.008171] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73a6f1e-ca20-4b3b-a9c7-ffe7dcef2d33 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.029888] env[63418]: DEBUG nova.compute.provider_tree [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.252505] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.252801] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.253010] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.254398] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.254668] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.254841] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.254990] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.255437] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.255437] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.255550] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.255707] env[63418]: DEBUG nova.virt.hardware [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.256689] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c17285a-ebc4-4414-b63a-98612cbfc5ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.265312] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488cb3b9-7f05-47a5-9107-4332380b3e11 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.270728] env[63418]: DEBUG nova.network.neutron [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.290359] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 545.302311] env[63418]: DEBUG oslo.service.loopingcall [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.302311] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 545.302515] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e0808e7-c5c0-48e7-83c9-796a729f88c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.321890] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 545.321890] env[63418]: value = "task-1244654" [ 545.321890] env[63418]: _type = "Task" [ 545.321890] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.333276] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244654, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.535029] env[63418]: DEBUG nova.scheduler.client.report [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 545.684539] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "f185a348-e91d-48d1-970b-473cc253cfdf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.684847] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "f185a348-e91d-48d1-970b-473cc253cfdf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.684948] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "f185a348-e91d-48d1-970b-473cc253cfdf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.685153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "f185a348-e91d-48d1-970b-473cc253cfdf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.685324] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "f185a348-e91d-48d1-970b-473cc253cfdf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.687824] env[63418]: INFO nova.compute.manager [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Terminating instance [ 545.773700] env[63418]: INFO nova.compute.manager [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] [instance: 8fbc105f-34f9-4ece-9e74-e473f310221a] Took 1.10 seconds to deallocate network for instance. [ 545.836032] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244654, 'name': CreateVM_Task, 'duration_secs': 0.401411} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.836257] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 545.836700] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.836857] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.837247] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 545.837532] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-121cab65-7e24-4814-8e8d-246181ff9362 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.843923] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 545.843923] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ad8ebe-bb29-7473-402e-a7ec43663bde" [ 545.843923] env[63418]: _type = "Task" [ 545.843923] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.852681] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ad8ebe-bb29-7473-402e-a7ec43663bde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.046103] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.046640] env[63418]: DEBUG nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 546.050025] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.682s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.052220] env[63418]: INFO nova.compute.claims [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.191812] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "refresh_cache-f185a348-e91d-48d1-970b-473cc253cfdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.192021] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "refresh_cache-f185a348-e91d-48d1-970b-473cc253cfdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.192265] env[63418]: DEBUG nova.network.neutron [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.355950] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ad8ebe-bb29-7473-402e-a7ec43663bde, 'name': SearchDatastore_Task, 'duration_secs': 0.040955} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.355950] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.357304] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 546.357304] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.357304] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.357304] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 546.357707] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10f8d43a-b003-4540-8a4e-77fb7b9f0c85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.368845] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 546.369069] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 546.369806] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-def6ee0d-fc3c-4af2-8044-60d52746d185 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.378512] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 546.378512] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5204715a-5208-2aef-c76c-2006a8e8cfb1" [ 546.378512] env[63418]: _type = "Task" [ 546.378512] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.392503] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5204715a-5208-2aef-c76c-2006a8e8cfb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.559124] env[63418]: DEBUG nova.compute.utils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.568337] env[63418]: DEBUG nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 546.736023] env[63418]: DEBUG nova.network.neutron [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.833171] env[63418]: INFO nova.scheduler.client.report [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Deleted allocations for instance 8fbc105f-34f9-4ece-9e74-e473f310221a [ 546.890088] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5204715a-5208-2aef-c76c-2006a8e8cfb1, 'name': SearchDatastore_Task, 'duration_secs': 0.019615} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.895716] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8456586-806d-42ca-b808-4e12e084404d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.904063] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 546.904063] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b96ef6-ffe1-a33d-101e-14dd69c62516" [ 546.904063] env[63418]: _type = "Task" [ 546.904063] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.915522] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b96ef6-ffe1-a33d-101e-14dd69c62516, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.959950] env[63418]: DEBUG nova.network.neutron [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.069292] env[63418]: DEBUG nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 547.342196] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aeb0e76-63f3-446a-b982-bc226f3720ee tempest-ImagesNegativeTestJSON-248159375 tempest-ImagesNegativeTestJSON-248159375-project-member] Lock "8fbc105f-34f9-4ece-9e74-e473f310221a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.524s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.387738] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1b8710-7af7-4361-a95c-e16ba782abb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.399076] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a4897a-4b32-4c69-a4af-319a5c05ac55 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.438659] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa6fb96-47ae-4a73-9658-c2942cbfe238 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.447972] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b96ef6-ffe1-a33d-101e-14dd69c62516, 'name': SearchDatastore_Task, 'duration_secs': 0.02713} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.447972] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.447972] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 547.447972] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14a3364f-182e-4b8e-9f83-dafaf5fa63e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.452768] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76efc10-1afd-4e67-b04d-167c0ad8c43b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.460777] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 547.460777] env[63418]: value = "task-1244656" [ 547.460777] env[63418]: _type = "Task" [ 547.460777] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.477611] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "refresh_cache-f185a348-e91d-48d1-970b-473cc253cfdf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.477611] env[63418]: DEBUG nova.compute.manager [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 547.477611] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.478148] env[63418]: DEBUG nova.compute.provider_tree [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.479771] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b10880-e8bb-4932-a776-d6a42bb82b3c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.492751] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.495088] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 547.495336] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-420ec3b9-ac0a-4954-92da-0829d97cf05b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.503561] env[63418]: DEBUG oslo_vmware.api [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 547.503561] env[63418]: value = "task-1244657" [ 547.503561] env[63418]: _type = "Task" [ 547.503561] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.513876] env[63418]: DEBUG oslo_vmware.api [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.844847] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 547.975859] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244656, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.987073] env[63418]: DEBUG nova.scheduler.client.report [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 548.025261] env[63418]: DEBUG oslo_vmware.api [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244657, 'name': PowerOffVM_Task, 'duration_secs': 0.152636} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.025540] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 548.025702] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 548.025959] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09cac791-234e-498c-9acd-8a3198b64a1a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.063392] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 548.063568] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 548.063797] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleting the datastore file [datastore2] f185a348-e91d-48d1-970b-473cc253cfdf {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 548.064108] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc2ce7cc-5dea-49f1-bf2a-80112c93a9b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.074303] env[63418]: DEBUG oslo_vmware.api [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 548.074303] env[63418]: value = "task-1244659" [ 548.074303] env[63418]: _type = "Task" [ 548.074303] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.083024] env[63418]: DEBUG oslo_vmware.api [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.091938] env[63418]: DEBUG nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 548.133407] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.133723] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.134145] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.134145] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.134601] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.134760] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.135306] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.135306] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.135306] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.135529] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.135600] env[63418]: DEBUG nova.virt.hardware [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.136498] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6c4304-3d29-4871-a536-f2a00c9a7d7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.151244] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c548a2-c7b8-4346-aac4-4f11409fd264 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.171392] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 548.181462] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Creating folder: Project (73c6513feb1b4a6e8ce8396893b88038). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 548.182022] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aadc2252-9983-44f2-83e6-d503ea208ba5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.196301] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Created folder: Project (73c6513feb1b4a6e8ce8396893b88038) in parent group-v268354. [ 548.196517] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Creating folder: Instances. Parent ref: group-v268369. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 548.196982] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a9e31b2-0fac-49cf-9ed7-8ac75d0d13b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.211909] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Created folder: Instances in parent group-v268369. [ 548.211909] env[63418]: DEBUG oslo.service.loopingcall [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.212207] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 548.212347] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e7b9106-fd42-405a-a2c7-c75c2d8f653f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.236237] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 548.236237] env[63418]: value = "task-1244662" [ 548.236237] env[63418]: _type = "Task" [ 548.236237] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.248267] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244662, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.384510] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.425794] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquiring lock "0c55e7b9-cb53-4525-8524-4ef62a51a89f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.426056] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Lock "0c55e7b9-cb53-4525-8524-4ef62a51a89f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.480263] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244656, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.493392] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.493392] env[63418]: DEBUG nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 548.499482] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.584s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.588765] env[63418]: DEBUG oslo_vmware.api [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212202} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.589268] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 548.591036] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 548.591036] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.591036] env[63418]: INFO nova.compute.manager [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Took 1.11 seconds to destroy the instance on the hypervisor. [ 548.591036] env[63418]: DEBUG oslo.service.loopingcall [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.591036] env[63418]: DEBUG nova.compute.manager [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 548.591036] env[63418]: DEBUG nova.network.neutron [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.656006] env[63418]: DEBUG nova.network.neutron [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.746790] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244662, 'name': CreateVM_Task, 'duration_secs': 0.380586} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.747094] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 548.747431] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.747735] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.747896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 548.748158] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73bc25cf-9135-4159-8455-0611a0205090 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.753868] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 548.753868] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521240c6-d289-68d9-f169-0e6956d58a4a" [ 548.753868] env[63418]: _type = "Task" [ 548.753868] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.762403] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521240c6-d289-68d9-f169-0e6956d58a4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.955157] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "584a7bc1-4651-49ff-a03c-94159f29f30f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.955157] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "584a7bc1-4651-49ff-a03c-94159f29f30f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.974454] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244656, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.002338] env[63418]: DEBUG nova.compute.utils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.007564] env[63418]: DEBUG nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 549.163343] env[63418]: DEBUG nova.network.neutron [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.275145] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521240c6-d289-68d9-f169-0e6956d58a4a, 'name': SearchDatastore_Task, 'duration_secs': 0.011611} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.275455] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.275678] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 549.275900] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.276047] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.276251] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 549.276511] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf39e094-b6e9-46bb-a94c-13ba3f12b886 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.287446] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 549.287617] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 549.288618] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9df9827-dfce-421a-8e7a-5306cd4bb140 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.302022] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 549.302022] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5247c7d8-5b8c-a45e-5e9d-87a2cf345b17" [ 549.302022] env[63418]: _type = "Task" [ 549.302022] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.312047] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5247c7d8-5b8c-a45e-5e9d-87a2cf345b17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.437798] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449c8c4d-bfb0-4f0d-895e-b779921dca39 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.449089] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1219daac-53b3-4ad0-b39c-fed6ee3bd598 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.486728] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2ae059-0965-4d85-a9c3-d06ef8a9d76d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.499050] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244656, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.647521} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.499256] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 549.499469] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 549.499837] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad05d4fe-0b71-4246-ac67-8fccc95057f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.502771] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12569912-ac2f-45ad-a2f1-9772c305c60e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.507687] env[63418]: DEBUG nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 549.520179] env[63418]: DEBUG nova.compute.provider_tree [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.527443] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 549.527443] env[63418]: value = "task-1244664" [ 549.527443] env[63418]: _type = "Task" [ 549.527443] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.533699] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244664, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.666633] env[63418]: INFO nova.compute.manager [-] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Took 1.07 seconds to deallocate network for instance. [ 549.813546] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5247c7d8-5b8c-a45e-5e9d-87a2cf345b17, 'name': SearchDatastore_Task, 'duration_secs': 0.01036} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.814631] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-297d0c27-9d3a-4b12-80a1-22077078126b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.825176] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 549.825176] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522e27bd-dc16-8cb8-fd39-d5620e666d72" [ 549.825176] env[63418]: _type = "Task" [ 549.825176] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.835839] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522e27bd-dc16-8cb8-fd39-d5620e666d72, 'name': SearchDatastore_Task, 'duration_secs': 0.010226} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.836111] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.836365] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 01deeb21-0e27-497f-9b85-c85949a3533d/01deeb21-0e27-497f-9b85-c85949a3533d.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 549.836617] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6d5dcfc-9827-4fe4-bd6f-46f2504d0b01 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.844276] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 549.844276] env[63418]: value = "task-1244665" [ 549.844276] env[63418]: _type = "Task" [ 549.844276] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.855405] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244665, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.949354] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "938304f0-1b30-4848-a82d-eb5acf732677" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.949640] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "938304f0-1b30-4848-a82d-eb5acf732677" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.024601] env[63418]: DEBUG nova.scheduler.client.report [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 550.043343] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244664, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06625} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.043635] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 550.044656] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea2dd86-8f4b-4667-9193-094271152cdb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.068868] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.068868] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eb8d073-a501-45ca-9be1-59710523ba26 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.093584] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 550.093584] env[63418]: value = "task-1244666" [ 550.093584] env[63418]: _type = "Task" [ 550.093584] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.106375] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244666, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.179641] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.357461] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244665, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.519956] env[63418]: DEBUG nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 550.539221] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.539719] env[63418]: ERROR nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Traceback (most recent call last): [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self.driver.spawn(context, instance, image_meta, [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] vm_ref = self.build_virtual_machine(instance, [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.539719] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] for vif in network_info: [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return self._sync_wrapper(fn, *args, **kwargs) [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self.wait() [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self[:] = self._gt.wait() [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return self._exit_event.wait() [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] result = hub.switch() [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.540130] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return self.greenlet.switch() [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] result = function(*args, **kwargs) [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] return func(*args, **kwargs) [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] raise e [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] nwinfo = self.network_api.allocate_for_instance( [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] created_port_ids = self._update_ports_for_instance( [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] with excutils.save_and_reraise_exception(): [ 550.540533] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] self.force_reraise() [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] raise self.value [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] updated_port = self._update_port( [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] _ensure_no_port_binding_failure(port) [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] raise exception.PortBindingFailed(port_id=port['id']) [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] nova.exception.PortBindingFailed: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. [ 550.540991] env[63418]: ERROR nova.compute.manager [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] [ 550.541325] env[63418]: DEBUG nova.compute.utils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 550.541701] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.935s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.543321] env[63418]: INFO nova.compute.claims [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.548626] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Build of instance 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0 was re-scheduled: Binding failed for port 1d10ead4-73ff-4946-bf33-e3e9a105b862, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 550.549128] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 550.549389] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquiring lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.549554] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Acquired lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.549711] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.568428] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.568736] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.568891] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.573310] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.573310] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.573310] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.573310] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.573310] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.573511] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.573511] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.573884] env[63418]: DEBUG nova.virt.hardware [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.575846] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98f2344-025a-43e4-939f-d603070cc5a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.585585] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93d4e9a-ea39-4aa1-b7e0-b1be3831861c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.622066] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 550.629029] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Creating folder: Project (408997fb3ce84317b38a39e978da621b). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 550.629725] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244666, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.632426] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e84ec88-e706-48e2-9c8b-76b60306fd2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.642994] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Created folder: Project (408997fb3ce84317b38a39e978da621b) in parent group-v268354. [ 550.643284] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Creating folder: Instances. Parent ref: group-v268372. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 550.643626] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c894e61-32e4-4c5a-8400-13fbb77c7976 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.659560] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Created folder: Instances in parent group-v268372. [ 550.659560] env[63418]: DEBUG oslo.service.loopingcall [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.659560] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 550.659560] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6fc4c0c-422f-4bb1-8035-c889f8cdb4d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.690056] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 550.690056] env[63418]: value = "task-1244669" [ 550.690056] env[63418]: _type = "Task" [ 550.690056] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.701024] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244669, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.860397] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244665, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515022} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.860666] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 01deeb21-0e27-497f-9b85-c85949a3533d/01deeb21-0e27-497f-9b85-c85949a3533d.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 550.860916] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 550.861208] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82759708-0d10-4112-bdea-d71e33afb840 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.871181] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 550.871181] env[63418]: value = "task-1244671" [ 550.871181] env[63418]: _type = "Task" [ 550.871181] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.883095] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244671, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.089401] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.112020] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244666, 'name': ReconfigVM_Task, 'duration_secs': 0.638854} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.112020] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Reconfigured VM instance instance-00000004 to attach disk [datastore1] ba67658a-668e-4fca-aefe-e838f7b05e2a/ba67658a-668e-4fca-aefe-e838f7b05e2a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.112020] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-931b6099-e721-4f53-8368-7d89a48a3f9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.120310] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 551.120310] env[63418]: value = "task-1244672" [ 551.120310] env[63418]: _type = "Task" [ 551.120310] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.133020] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244672, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.207023] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244669, 'name': CreateVM_Task, 'duration_secs': 0.35632} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.207538] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 551.207797] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.208054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.208273] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 551.208515] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae920ffc-c1d0-4da6-b003-3495ae96b67c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.214799] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 551.214799] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5272bcd6-cabf-0acf-439c-021ee20009fe" [ 551.214799] env[63418]: _type = "Task" [ 551.214799] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.219154] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.230080] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5272bcd6-cabf-0acf-439c-021ee20009fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.347863] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "76e02181-1ccd-43f9-9aaa-cf1918691705" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.347954] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "76e02181-1ccd-43f9-9aaa-cf1918691705" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.382736] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244671, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166837} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.384557] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 551.386975] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42de6647-efa9-49bb-8f8c-b201005c1144 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.416347] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] 01deeb21-0e27-497f-9b85-c85949a3533d/01deeb21-0e27-497f-9b85-c85949a3533d.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 551.416347] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abc7f1b1-91e9-460a-8af8-ca8d009a290c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.442380] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 551.442380] env[63418]: value = "task-1244673" [ 551.442380] env[63418]: _type = "Task" [ 551.442380] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.451773] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.634532] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244672, 'name': Rename_Task, 'duration_secs': 0.173638} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.638037] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 551.638037] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36b0690f-3d2b-4ec6-8ab7-509b357e96f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.646118] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Waiting for the task: (returnval){ [ 551.646118] env[63418]: value = "task-1244674" [ 551.646118] env[63418]: _type = "Task" [ 551.646118] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.659172] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244674, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.726391] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Releasing lock "refresh_cache-9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.726603] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 551.728026] env[63418]: DEBUG nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 551.728026] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 551.738454] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5272bcd6-cabf-0acf-439c-021ee20009fe, 'name': SearchDatastore_Task, 'duration_secs': 0.012662} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.738924] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.739200] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 551.739436] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.739575] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.739749] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 551.740033] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74e61c22-7405-408a-8bb6-60fb7e117d89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.753021] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 551.754364] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 551.755392] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ba505bf-f4ce-4600-9331-52bf037beb04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.768241] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 551.768241] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526e81ed-24bc-58c8-3db2-8635c6acad5e" [ 551.768241] env[63418]: _type = "Task" [ 551.768241] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.780514] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526e81ed-24bc-58c8-3db2-8635c6acad5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.795472] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.955329] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244673, 'name': ReconfigVM_Task, 'duration_secs': 0.33789} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.955665] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Reconfigured VM instance instance-0000000b to attach disk [datastore1] 01deeb21-0e27-497f-9b85-c85949a3533d/01deeb21-0e27-497f-9b85-c85949a3533d.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.958976] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75dd9174-3d90-49ec-8104-c3aa2ff65ef2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.969091] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 551.969091] env[63418]: value = "task-1244675" [ 551.969091] env[63418]: _type = "Task" [ 551.969091] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.986986] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244675, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.043198] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "b56238f3-3507-49e0-8630-5f5a093c1101" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.043198] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "b56238f3-3507-49e0-8630-5f5a093c1101" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.066101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1217a6ba-86c7-4f57-9f90-2601f73f0364 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.081572] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5becfc-d4fa-4899-9a6c-78aa2c48226a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.115146] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d80f8f-264a-4599-96a5-e489dbb31b67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.124630] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb41d21-7914-4387-806a-2937c26c1cfb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.141116] env[63418]: DEBUG nova.compute.provider_tree [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.157295] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244674, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.286822] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526e81ed-24bc-58c8-3db2-8635c6acad5e, 'name': SearchDatastore_Task, 'duration_secs': 0.019418} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.286822] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b35036f6-3f46-4f60-92ab-a749e5f4f9ca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.295518] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 552.295518] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520568d4-fb8a-9388-6ca1-774403175f6c" [ 552.295518] env[63418]: _type = "Task" [ 552.295518] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.301893] env[63418]: DEBUG nova.network.neutron [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.315429] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520568d4-fb8a-9388-6ca1-774403175f6c, 'name': SearchDatastore_Task, 'duration_secs': 0.010527} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.316324] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.316583] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] ed637299-7e69-43d7-85f9-8fcbd6e90dec/ed637299-7e69-43d7-85f9-8fcbd6e90dec.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 552.316839] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5cfe6f0d-19db-4096-913a-c0202eb969f5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.326490] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 552.326490] env[63418]: value = "task-1244676" [ 552.326490] env[63418]: _type = "Task" [ 552.326490] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.337419] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244676, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.486396] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244675, 'name': Rename_Task, 'duration_secs': 0.177951} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.486734] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 552.486976] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92ec2556-65aa-41e4-b168-9881943a1664 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.495982] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 552.495982] env[63418]: value = "task-1244677" [ 552.495982] env[63418]: _type = "Task" [ 552.495982] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.508467] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.645026] env[63418]: DEBUG nova.scheduler.client.report [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 552.663488] env[63418]: DEBUG oslo_vmware.api [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Task: {'id': task-1244674, 'name': PowerOnVM_Task, 'duration_secs': 0.674178} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.663488] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 552.663488] env[63418]: DEBUG nova.compute.manager [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 552.663712] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62cc2122-f6ac-47de-8c49-fbf036c8ffd5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.807637] env[63418]: INFO nova.compute.manager [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] [instance: 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0] Took 1.08 seconds to deallocate network for instance. [ 552.843622] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244676, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.019719] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244677, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.157415] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.616s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.157949] env[63418]: DEBUG nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 553.162286] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.571s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.162661] env[63418]: INFO nova.compute.claims [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 553.184735] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.341299] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244676, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.809581} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.341592] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] ed637299-7e69-43d7-85f9-8fcbd6e90dec/ed637299-7e69-43d7-85f9-8fcbd6e90dec.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 553.342809] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 553.342809] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86a5df0d-7af1-4e0f-b541-abfc7e01b984 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.356321] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 553.356321] env[63418]: value = "task-1244678" [ 553.356321] env[63418]: _type = "Task" [ 553.356321] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.368731] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244678, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.509138] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244677, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.670058] env[63418]: DEBUG nova.compute.utils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.672032] env[63418]: DEBUG nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 553.856665] env[63418]: INFO nova.scheduler.client.report [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Deleted allocations for instance 9a70a79a-64a9-4de3-9d29-ada5d43f4dc0 [ 553.879576] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244678, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.223691} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.883072] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 553.884063] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2872113-bd7c-4159-b9da-b590ff586024 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.920098] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] ed637299-7e69-43d7-85f9-8fcbd6e90dec/ed637299-7e69-43d7-85f9-8fcbd6e90dec.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 553.922259] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bb11bcd-627e-47a4-ae0c-58464f10f81a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.944452] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 553.944452] env[63418]: value = "task-1244679" [ 553.944452] env[63418]: _type = "Task" [ 553.944452] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.954902] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244679, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.012460] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244677, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.175363] env[63418]: DEBUG nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 554.369536] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1e8db754-0d50-4e33-9d01-697eb19f5b20 tempest-ServerExternalEventsTest-1464480875 tempest-ServerExternalEventsTest-1464480875-project-member] Lock "9a70a79a-64a9-4de3-9d29-ada5d43f4dc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.207s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.465277] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244679, 'name': ReconfigVM_Task, 'duration_secs': 0.312126} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.465639] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Reconfigured VM instance instance-0000000c to attach disk [datastore2] ed637299-7e69-43d7-85f9-8fcbd6e90dec/ed637299-7e69-43d7-85f9-8fcbd6e90dec.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 554.468684] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85b3acd5-2ae2-4650-97c6-e0ecfc200317 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.485598] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 554.485598] env[63418]: value = "task-1244680" [ 554.485598] env[63418]: _type = "Task" [ 554.485598] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.501223] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244680, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.513425] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244677, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.559136] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquiring lock "59442c77-234c-48c2-872d-85ed9bd8cb80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.559455] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Lock "59442c77-234c-48c2-872d-85ed9bd8cb80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.700474] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52db4689-e140-467d-899b-f4d3353cf50c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.718240] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b8d66f-0287-47e4-b844-877159660579 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.759724] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dfcecb-2e4d-4bf1-bd4e-11895537a26c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.772354] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a1d16d-af47-407a-9463-2cb2b6e2df17 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.790679] env[63418]: DEBUG nova.compute.provider_tree [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.872630] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 554.997137] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244680, 'name': Rename_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.011287] env[63418]: DEBUG oslo_vmware.api [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244677, 'name': PowerOnVM_Task, 'duration_secs': 2.059827} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.011656] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 555.012342] env[63418]: INFO nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Took 6.92 seconds to spawn the instance on the hypervisor. [ 555.012759] env[63418]: DEBUG nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 555.013676] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d29d4e-4073-433e-9bd6-7136630dbc8b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.188966] env[63418]: DEBUG nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 555.203104] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquiring lock "b24d095a-6814-413b-bf89-511e22df740f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.203684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Lock "b24d095a-6814-413b-bf89-511e22df740f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.219726] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.219984] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.220349] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.220349] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.221569] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.221808] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.221992] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.222177] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.222350] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.223045] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.223045] env[63418]: DEBUG nova.virt.hardware [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.223953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35d1f11-a237-466a-b77a-46c13199b4ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.235013] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62198d2d-c641-4443-9fa6-68034803e60b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.253471] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 555.260600] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Creating folder: Project (55865e13dfb148bb894f5b8593723e25). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 555.261027] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3444327-92c2-4359-9f0f-189359615145 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.274210] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Created folder: Project (55865e13dfb148bb894f5b8593723e25) in parent group-v268354. [ 555.274438] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Creating folder: Instances. Parent ref: group-v268375. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 555.274692] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4f474d5-f1d5-40fe-8b44-f83ebdc2098f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.285877] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Created folder: Instances in parent group-v268375. [ 555.286325] env[63418]: DEBUG oslo.service.loopingcall [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.286679] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 555.286920] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00bbff36-ef6d-4902-ae9e-03ef8768648a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.301856] env[63418]: DEBUG nova.scheduler.client.report [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 555.312469] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 555.312469] env[63418]: value = "task-1244683" [ 555.312469] env[63418]: _type = "Task" [ 555.312469] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.326112] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244683, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.400557] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.496048] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244680, 'name': Rename_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.533638] env[63418]: INFO nova.compute.manager [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Took 29.45 seconds to build instance. [ 555.806915] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.646s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.807532] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 555.811198] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.026s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.813937] env[63418]: INFO nova.compute.claims [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.834642] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244683, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.999817] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244680, 'name': Rename_Task, 'duration_secs': 1.167426} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.000266] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 556.000601] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b98d5524-20f9-485b-88b5-29fc0593e5f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.011794] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 556.011794] env[63418]: value = "task-1244684" [ 556.011794] env[63418]: _type = "Task" [ 556.011794] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.026860] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.036830] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f522477-c95d-43f6-9570-2d272f47dc4b tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "01deeb21-0e27-497f-9b85-c85949a3533d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.399s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.252105] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquiring lock "601f72f4-05bc-4d73-9b57-612cad2d8c62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.253245] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Lock "601f72f4-05bc-4d73-9b57-612cad2d8c62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.319141] env[63418]: DEBUG nova.compute.utils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 556.324797] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 556.324982] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 556.340021] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244683, 'name': CreateVM_Task, 'duration_secs': 0.62276} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.340358] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 556.340887] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.341213] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.341793] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 556.342384] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ab83c5a-49c0-40f8-afab-e22ba7d31a19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.348568] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 556.348568] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d5a8fb-a5dd-c5a9-04ab-941c4aef388e" [ 556.348568] env[63418]: _type = "Task" [ 556.348568] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.364144] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d5a8fb-a5dd-c5a9-04ab-941c4aef388e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.449401] env[63418]: DEBUG nova.policy [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59ba7218583446b896fab51c87a6d3dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c542b46f8824ee7897f8b5ece368978', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 556.524703] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244684, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.542837] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 556.674237] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "6d47c52c-c1fb-4ade-a3ba-dc393716eb79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.674237] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "6d47c52c-c1fb-4ade-a3ba-dc393716eb79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.826354] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 556.863752] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d5a8fb-a5dd-c5a9-04ab-941c4aef388e, 'name': SearchDatastore_Task, 'duration_secs': 0.013006} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.867357] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.868014] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 556.868014] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.868014] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.868194] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 556.870478] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb9ffee7-0026-40ae-ab86-e2b7642d59aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.881449] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 556.882627] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 556.883568] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6deceae-2807-4a01-a989-0be344e32c2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.892804] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 556.892804] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52697c8a-a53a-1461-234a-012776ffee65" [ 556.892804] env[63418]: _type = "Task" [ 556.892804] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.910858] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52697c8a-a53a-1461-234a-012776ffee65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.023082] env[63418]: DEBUG oslo_vmware.api [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244684, 'name': PowerOnVM_Task, 'duration_secs': 0.612332} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.023374] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 557.023572] env[63418]: INFO nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Took 6.50 seconds to spawn the instance on the hypervisor. [ 557.023746] env[63418]: DEBUG nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 557.024580] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64eca49b-03fa-411f-af01-0bfd24d9ac3b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.063189] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.405703] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52697c8a-a53a-1461-234a-012776ffee65, 'name': SearchDatastore_Task, 'duration_secs': 0.019449} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.406688] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-005e791e-835e-4a1b-bbc4-111e5936d199 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.413676] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 557.413676] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52861d9a-c592-e137-705e-53d5aae51ae4" [ 557.413676] env[63418]: _type = "Task" [ 557.413676] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.426264] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "ba67658a-668e-4fca-aefe-e838f7b05e2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.429209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "ba67658a-668e-4fca-aefe-e838f7b05e2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.429209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "ba67658a-668e-4fca-aefe-e838f7b05e2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.429209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "ba67658a-668e-4fca-aefe-e838f7b05e2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.429767] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "ba67658a-668e-4fca-aefe-e838f7b05e2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.433811] env[63418]: INFO nova.compute.manager [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Terminating instance [ 557.442149] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52861d9a-c592-e137-705e-53d5aae51ae4, 'name': SearchDatastore_Task, 'duration_secs': 0.0116} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.443597] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.443951] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 557.446025] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67c765e2-e570-4021-93ef-c4c2516edf79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.454738] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 557.454738] env[63418]: value = "task-1244685" [ 557.454738] env[63418]: _type = "Task" [ 557.454738] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.456982] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee87089-335c-4dde-92f4-1a55ebe9ddfd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.466074] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Successfully created port: 2163b943-c087-4595-af3d-2f900fff8221 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 557.483027] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquiring lock "c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.483027] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Lock "c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.483219] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.484755] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d98bf6-b8be-49f0-b019-10c24148b0d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.520587] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3070be80-3c4c-4576-a5a3-9dc55f6271a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.531267] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d5fb0b-43a4-47d8-8000-d94f7d01f79a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.554461] env[63418]: DEBUG nova.compute.provider_tree [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.555796] env[63418]: INFO nova.compute.manager [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Took 30.20 seconds to build instance. [ 557.847835] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 557.885769] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.886982] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.886982] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.886982] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.886982] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.886982] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.887309] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.887309] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.887309] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.887457] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.887643] env[63418]: DEBUG nova.virt.hardware [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.888851] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dfd947-7250-4ee5-b110-8d1ba03997c2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.899031] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222b2c7a-b8a1-4b80-9e09-3482987f66e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.946692] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "refresh_cache-ba67658a-668e-4fca-aefe-e838f7b05e2a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.946692] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquired lock "refresh_cache-ba67658a-668e-4fca-aefe-e838f7b05e2a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.946692] env[63418]: DEBUG nova.network.neutron [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.968783] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244685, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.056822] env[63418]: DEBUG nova.scheduler.client.report [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 558.060195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-cec9448d-72b7-4b0b-bcf5-3634c4483e43 tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.961s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.471113] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577776} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.472187] env[63418]: DEBUG nova.network.neutron [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.474309] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 558.474426] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 558.474759] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c2e019f-e1f3-406d-a6ca-8f36ce0fb5cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.484102] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 558.484102] env[63418]: value = "task-1244686" [ 558.484102] env[63418]: _type = "Task" [ 558.484102] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.494610] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.562712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.751s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.564120] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 558.571831] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 558.581520] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.054s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.595035] env[63418]: DEBUG nova.network.neutron [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.841800] env[63418]: DEBUG nova.compute.manager [None req-93ff0c1e-8b74-433a-9305-690d6e08029b tempest-ServerDiagnosticsV248Test-1168567041 tempest-ServerDiagnosticsV248Test-1168567041-project-admin] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 558.847019] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e83c9a-ce85-4afb-99c5-ee2e02b0b5f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.853294] env[63418]: INFO nova.compute.manager [None req-93ff0c1e-8b74-433a-9305-690d6e08029b tempest-ServerDiagnosticsV248Test-1168567041 tempest-ServerDiagnosticsV248Test-1168567041-project-admin] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Retrieving diagnostics [ 558.854556] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c52f519-ed03-49ed-9c4d-d5dcdcbfd7e2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.997982] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.221057} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.998331] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 558.999161] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749a27ab-5b13-4598-a983-6621d7367c9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.026911] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 559.027245] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82874bf7-fb1b-4ab0-af11-a3e5ea2f7bb8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.057458] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 559.057458] env[63418]: value = "task-1244687" [ 559.057458] env[63418]: _type = "Task" [ 559.057458] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.068271] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244687, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.077162] env[63418]: DEBUG nova.compute.utils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.078615] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 559.078770] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 559.098881] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Releasing lock "refresh_cache-ba67658a-668e-4fca-aefe-e838f7b05e2a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.099411] env[63418]: DEBUG nova.compute.manager [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 559.099504] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.100391] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf831f8-fbf4-4548-8d92-7f90982906a1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.109296] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 559.109569] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34e86f59-d6fc-4b0c-aa86-e375222ad01b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.112636] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.119121] env[63418]: DEBUG oslo_vmware.api [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 559.119121] env[63418]: value = "task-1244688" [ 559.119121] env[63418]: _type = "Task" [ 559.119121] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.129388] env[63418]: DEBUG oslo_vmware.api [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.166816] env[63418]: DEBUG nova.policy [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5463354b2304ec0bd7cdce3a165412c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e822358578e0405ba691dd88f17f2088', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.266324] env[63418]: ERROR nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 559.266324] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.266324] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.266324] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.266324] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.266324] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.266324] env[63418]: ERROR nova.compute.manager raise self.value [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.266324] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.266324] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.266324] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.266899] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.266899] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.266899] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 559.266899] env[63418]: ERROR nova.compute.manager [ 559.266899] env[63418]: Traceback (most recent call last): [ 559.266899] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.266899] env[63418]: listener.cb(fileno) [ 559.266899] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.266899] env[63418]: result = function(*args, **kwargs) [ 559.266899] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.266899] env[63418]: return func(*args, **kwargs) [ 559.266899] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 559.266899] env[63418]: raise e [ 559.266899] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.266899] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 559.266899] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.266899] env[63418]: created_port_ids = self._update_ports_for_instance( [ 559.266899] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.266899] env[63418]: with excutils.save_and_reraise_exception(): [ 559.266899] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.266899] env[63418]: self.force_reraise() [ 559.266899] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.266899] env[63418]: raise self.value [ 559.266899] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.266899] env[63418]: updated_port = self._update_port( [ 559.266899] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.266899] env[63418]: _ensure_no_port_binding_failure(port) [ 559.266899] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.266899] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.268400] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 559.268400] env[63418]: Removing descriptor: 16 [ 559.268400] env[63418]: ERROR nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] Traceback (most recent call last): [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] yield resources [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self.driver.spawn(context, instance, image_meta, [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.268400] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] vm_ref = self.build_virtual_machine(instance, [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] for vif in network_info: [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return self._sync_wrapper(fn, *args, **kwargs) [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self.wait() [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self[:] = self._gt.wait() [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return self._exit_event.wait() [ 559.269897] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] result = hub.switch() [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return self.greenlet.switch() [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] result = function(*args, **kwargs) [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return func(*args, **kwargs) [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] raise e [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] nwinfo = self.network_api.allocate_for_instance( [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.270584] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] created_port_ids = self._update_ports_for_instance( [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] with excutils.save_and_reraise_exception(): [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self.force_reraise() [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] raise self.value [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] updated_port = self._update_port( [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] _ensure_no_port_binding_failure(port) [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.271131] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] raise exception.PortBindingFailed(port_id=port['id']) [ 559.272597] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 559.272597] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] [ 559.272597] env[63418]: INFO nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Terminating instance [ 559.556957] env[63418]: DEBUG nova.compute.manager [None req-8602b4a3-0bee-44e1-9769-123ec17ca0a6 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 559.557920] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06695c2-452a-45c2-976e-6afc975e3f82 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.586722] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.586722] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 559.590274] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02d1045-a4ff-43bd-95e1-87316e10ec69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.600779] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df40b1eb-d0eb-463b-aeb4-75597e412535 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.639433] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa50cbaf-c37a-4c04-8a86-b1f19129287f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.655929] env[63418]: DEBUG oslo_vmware.api [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244688, 'name': PowerOffVM_Task, 'duration_secs': 0.251658} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.658157] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98918e67-8fcc-4d5a-afc8-b13811f9a55c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.662816] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 559.663056] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 559.663604] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d42eec2a-9ace-4cb1-a1a5-233782284bc1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.668394] env[63418]: DEBUG nova.compute.manager [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Received event network-changed-2163b943-c087-4595-af3d-2f900fff8221 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 559.668394] env[63418]: DEBUG nova.compute.manager [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Refreshing instance network info cache due to event network-changed-2163b943-c087-4595-af3d-2f900fff8221. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 559.668394] env[63418]: DEBUG oslo_concurrency.lockutils [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] Acquiring lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.668394] env[63418]: DEBUG oslo_concurrency.lockutils [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] Acquired lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.668584] env[63418]: DEBUG nova.network.neutron [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Refreshing network info cache for port 2163b943-c087-4595-af3d-2f900fff8221 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 559.679177] env[63418]: DEBUG nova.compute.provider_tree [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.686731] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Successfully created port: 565d8ffd-fbd2-45c3-918f-51b7f9b07557 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.711554] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 559.711844] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 559.712101] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleting the datastore file [datastore1] ba67658a-668e-4fca-aefe-e838f7b05e2a {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 559.712456] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4c32ab6-8618-4a7d-8d3a-d3d73b9efa54 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.725116] env[63418]: DEBUG oslo_vmware.api [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for the task: (returnval){ [ 559.725116] env[63418]: value = "task-1244690" [ 559.725116] env[63418]: _type = "Task" [ 559.725116] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.730994] env[63418]: DEBUG nova.network.neutron [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.739496] env[63418]: DEBUG oslo_vmware.api [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.771422] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquiring lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.029366] env[63418]: DEBUG nova.network.neutron [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.067911] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.086229] env[63418]: INFO nova.compute.manager [None req-8602b4a3-0bee-44e1-9769-123ec17ca0a6 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] instance snapshotting [ 560.087978] env[63418]: DEBUG nova.objects.instance [None req-8602b4a3-0bee-44e1-9769-123ec17ca0a6 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lazy-loading 'flavor' on Instance uuid 01deeb21-0e27-497f-9b85-c85949a3533d {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 560.188488] env[63418]: DEBUG nova.scheduler.client.report [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 560.236632] env[63418]: DEBUG oslo_vmware.api [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Task: {'id': task-1244690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.422291} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.236882] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 560.237174] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 560.237266] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.237480] env[63418]: INFO nova.compute.manager [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 560.237811] env[63418]: DEBUG oslo.service.loopingcall [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.237961] env[63418]: DEBUG nova.compute.manager [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 560.238072] env[63418]: DEBUG nova.network.neutron [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.286367] env[63418]: DEBUG nova.network.neutron [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.511290] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "01deeb21-0e27-497f-9b85-c85949a3533d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.512118] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "01deeb21-0e27-497f-9b85-c85949a3533d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.512118] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "01deeb21-0e27-497f-9b85-c85949a3533d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.512118] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "01deeb21-0e27-497f-9b85-c85949a3533d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.512118] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "01deeb21-0e27-497f-9b85-c85949a3533d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.514341] env[63418]: INFO nova.compute.manager [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Terminating instance [ 560.534038] env[63418]: DEBUG oslo_concurrency.lockutils [req-6eb452c7-9b64-486c-9faf-fc028b5c111b req-cfef90d0-7b9c-475a-97b0-641900684f1e service nova] Releasing lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.534038] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquired lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.534038] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.572831] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244687, 'name': ReconfigVM_Task, 'duration_secs': 1.363961} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.573256] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Reconfigured VM instance instance-0000000d to attach disk [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 560.573968] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ec48af7-3cce-444a-a1f6-a2d6bfb8474d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.586017] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 560.586017] env[63418]: value = "task-1244691" [ 560.586017] env[63418]: _type = "Task" [ 560.586017] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.595509] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244691, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.597527] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d0e8a7-f940-4326-96c4-e7a14ba6151d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.618823] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 560.622450] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a3dd6a-b22a-45cd-b0f7-56d257ffa654 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.661802] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.663019] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.663019] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.663156] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.663213] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.663356] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.663579] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.663732] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.663892] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.664078] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.664253] env[63418]: DEBUG nova.virt.hardware [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.665472] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e163c37c-bfe4-451f-97f3-851ca6cf64d5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.677558] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0f6d66-1eb9-4562-a68c-f76b99604423 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.707033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.127s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.707033] env[63418]: ERROR nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Traceback (most recent call last): [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self.driver.spawn(context, instance, image_meta, [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.707033] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] vm_ref = self.build_virtual_machine(instance, [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] for vif in network_info: [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return self._sync_wrapper(fn, *args, **kwargs) [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self.wait() [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self[:] = self._gt.wait() [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return self._exit_event.wait() [ 560.707372] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] result = hub.switch() [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return self.greenlet.switch() [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] result = function(*args, **kwargs) [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] return func(*args, **kwargs) [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] raise e [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] nwinfo = self.network_api.allocate_for_instance( [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.707788] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] created_port_ids = self._update_ports_for_instance( [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] with excutils.save_and_reraise_exception(): [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] self.force_reraise() [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] raise self.value [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] updated_port = self._update_port( [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] _ensure_no_port_binding_failure(port) [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.708184] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] raise exception.PortBindingFailed(port_id=port['id']) [ 560.708539] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] nova.exception.PortBindingFailed: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. [ 560.708539] env[63418]: ERROR nova.compute.manager [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] [ 560.710438] env[63418]: DEBUG nova.compute.utils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.712520] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.105s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.714112] env[63418]: INFO nova.compute.claims [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.717455] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Build of instance d255e7af-b50d-4b30-8bec-9e2ee2de2c80 was re-scheduled: Binding failed for port 36803e73-f0ee-4ef3-b808-d0f51174734c, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 560.718125] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 560.718481] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.718960] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquired lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.719243] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.790456] env[63418]: DEBUG nova.network.neutron [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.019070] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "refresh_cache-01deeb21-0e27-497f-9b85-c85949a3533d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.019251] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquired lock "refresh_cache-01deeb21-0e27-497f-9b85-c85949a3533d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.019419] env[63418]: DEBUG nova.network.neutron [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.072128] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.097018] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244691, 'name': Rename_Task, 'duration_secs': 0.338189} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.097018] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 561.097192] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6308a211-41ca-4ff9-8a5d-97c2232d9ee9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.106074] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 561.106074] env[63418]: value = "task-1244692" [ 561.106074] env[63418]: _type = "Task" [ 561.106074] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.116787] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.135703] env[63418]: DEBUG nova.compute.manager [None req-8602b4a3-0bee-44e1-9769-123ec17ca0a6 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Instance disappeared during snapshot {{(pid=63418) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 561.263168] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.293364] env[63418]: INFO nova.compute.manager [-] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Took 1.06 seconds to deallocate network for instance. [ 561.308078] env[63418]: DEBUG nova.compute.manager [None req-8602b4a3-0bee-44e1-9769-123ec17ca0a6 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Found 0 images (rotation: 2) {{(pid=63418) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 561.317813] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.534606] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.558118] env[63418]: DEBUG nova.network.neutron [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.624050] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244692, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.738027] env[63418]: DEBUG nova.network.neutron [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.806381] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.821551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Releasing lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.822534] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 561.822534] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.822534] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6916e220-c2e9-4653-9dbf-02647d6baca2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.837715] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54c751d-2df7-4f52-af96-ef48d639db5f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.872691] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f714b41e-b537-4feb-bd71-53b5db089485 could not be found. [ 561.873046] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 561.873270] env[63418]: INFO nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Took 0.05 seconds to destroy the instance on the hypervisor. [ 561.873554] env[63418]: DEBUG oslo.service.loopingcall [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.876622] env[63418]: DEBUG nova.compute.manager [-] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 561.876737] env[63418]: DEBUG nova.network.neutron [-] [instance: f714b41e-b537-4feb-bd71-53b5db089485] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.903377] env[63418]: ERROR nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 561.903377] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.903377] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.903377] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.903377] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.903377] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.903377] env[63418]: ERROR nova.compute.manager raise self.value [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.903377] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.903377] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.903377] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.903938] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.903938] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.903938] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 561.903938] env[63418]: ERROR nova.compute.manager [ 561.903938] env[63418]: Traceback (most recent call last): [ 561.903938] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.903938] env[63418]: listener.cb(fileno) [ 561.903938] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.903938] env[63418]: result = function(*args, **kwargs) [ 561.903938] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.903938] env[63418]: return func(*args, **kwargs) [ 561.903938] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.903938] env[63418]: raise e [ 561.903938] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.903938] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 561.903938] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.903938] env[63418]: created_port_ids = self._update_ports_for_instance( [ 561.903938] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.903938] env[63418]: with excutils.save_and_reraise_exception(): [ 561.903938] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.903938] env[63418]: self.force_reraise() [ 561.903938] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.903938] env[63418]: raise self.value [ 561.903938] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.903938] env[63418]: updated_port = self._update_port( [ 561.903938] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.903938] env[63418]: _ensure_no_port_binding_failure(port) [ 561.903938] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.903938] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.904882] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 561.904882] env[63418]: Removing descriptor: 15 [ 561.904882] env[63418]: ERROR nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Traceback (most recent call last): [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] yield resources [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self.driver.spawn(context, instance, image_meta, [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.904882] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] vm_ref = self.build_virtual_machine(instance, [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] for vif in network_info: [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return self._sync_wrapper(fn, *args, **kwargs) [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self.wait() [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self[:] = self._gt.wait() [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return self._exit_event.wait() [ 561.905278] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] result = hub.switch() [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return self.greenlet.switch() [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] result = function(*args, **kwargs) [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return func(*args, **kwargs) [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] raise e [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] nwinfo = self.network_api.allocate_for_instance( [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.905674] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] created_port_ids = self._update_ports_for_instance( [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] with excutils.save_and_reraise_exception(): [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self.force_reraise() [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] raise self.value [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] updated_port = self._update_port( [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] _ensure_no_port_binding_failure(port) [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.906128] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] raise exception.PortBindingFailed(port_id=port['id']) [ 561.906510] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 561.906510] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] [ 561.906510] env[63418]: INFO nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Terminating instance [ 561.930233] env[63418]: DEBUG nova.network.neutron [-] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.042657] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Releasing lock "refresh_cache-d255e7af-b50d-4b30-8bec-9e2ee2de2c80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.042657] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 562.042657] env[63418]: DEBUG nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 562.042657] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.079969] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.127440] env[63418]: DEBUG oslo_vmware.api [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244692, 'name': PowerOnVM_Task, 'duration_secs': 0.813978} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.127713] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 562.127913] env[63418]: INFO nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Took 6.94 seconds to spawn the instance on the hypervisor. [ 562.130812] env[63418]: DEBUG nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 562.131326] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec2852c-2bb5-4bf0-9a0b-18185a6b66ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.164365] env[63418]: DEBUG nova.compute.manager [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Received event network-changed-565d8ffd-fbd2-45c3-918f-51b7f9b07557 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 562.164365] env[63418]: DEBUG nova.compute.manager [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Refreshing instance network info cache due to event network-changed-565d8ffd-fbd2-45c3-918f-51b7f9b07557. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 562.164365] env[63418]: DEBUG oslo_concurrency.lockutils [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] Acquiring lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.164365] env[63418]: DEBUG oslo_concurrency.lockutils [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] Acquired lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.164365] env[63418]: DEBUG nova.network.neutron [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Refreshing network info cache for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 562.231377] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquiring lock "f6d62d1b-0333-40ac-92e8-a798e58d495c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.231377] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Lock "f6d62d1b-0333-40ac-92e8-a798e58d495c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.250595] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Releasing lock "refresh_cache-01deeb21-0e27-497f-9b85-c85949a3533d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.252016] env[63418]: DEBUG nova.compute.manager [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 562.252016] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.252576] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5383a02f-34ae-4ae1-bf38-499d0cc116b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.265022] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 562.265863] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66f9cf9f-3e50-4fb0-9c71-cb30a7db46ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.276203] env[63418]: DEBUG oslo_vmware.api [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 562.276203] env[63418]: value = "task-1244693" [ 562.276203] env[63418]: _type = "Task" [ 562.276203] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.291254] env[63418]: DEBUG oslo_vmware.api [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.328464] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529a5f27-aaf0-4716-bd4d-ae6505f6db73 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.338671] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbf2297-ec5e-44c0-be6f-06c0ebb5d457 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.381224] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf8da57-27cc-4903-b22e-0b2915a6f5a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.391344] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6404a6e-06f8-47b4-9280-b9b98c07931f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.399010] env[63418]: DEBUG nova.compute.manager [req-536edcea-2cee-47d5-825f-40fdeca240f5 req-df7d79c5-30c7-4eca-8ced-b6d80c10fe31 service nova] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Received event network-vif-deleted-2163b943-c087-4595-af3d-2f900fff8221 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 562.414379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquiring lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.414647] env[63418]: DEBUG nova.compute.provider_tree [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.435154] env[63418]: DEBUG nova.network.neutron [-] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.585801] env[63418]: DEBUG nova.network.neutron [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.663885] env[63418]: INFO nova.compute.manager [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Took 31.08 seconds to build instance. [ 562.702711] env[63418]: DEBUG nova.network.neutron [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.787397] env[63418]: DEBUG oslo_vmware.api [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244693, 'name': PowerOffVM_Task, 'duration_secs': 0.420736} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.789225] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 562.789225] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 562.789225] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb79df32-6d86-4145-9d18-5498e00c1565 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.822145] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 562.822242] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 562.822481] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Deleting the datastore file [datastore1] 01deeb21-0e27-497f-9b85-c85949a3533d {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 562.822669] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a46bb868-535f-4a1a-8755-a9f4559e76cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.837487] env[63418]: DEBUG oslo_vmware.api [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for the task: (returnval){ [ 562.837487] env[63418]: value = "task-1244695" [ 562.837487] env[63418]: _type = "Task" [ 562.837487] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.846919] env[63418]: DEBUG oslo_vmware.api [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.898615] env[63418]: DEBUG nova.network.neutron [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.917360] env[63418]: DEBUG nova.scheduler.client.report [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 562.939070] env[63418]: INFO nova.compute.manager [-] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Took 1.06 seconds to deallocate network for instance. [ 562.942651] env[63418]: DEBUG nova.compute.claims [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.942979] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.088812] env[63418]: INFO nova.compute.manager [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: d255e7af-b50d-4b30-8bec-9e2ee2de2c80] Took 1.05 seconds to deallocate network for instance. [ 563.169656] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50bfcade-be5c-4f8d-9499-4fc34123f133 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "a8c7f192-4672-43cc-8c38-0c33ce633765" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.444s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.352020] env[63418]: DEBUG oslo_vmware.api [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Task: {'id': task-1244695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187045} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.352020] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 563.352439] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 563.354123] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.354123] env[63418]: INFO nova.compute.manager [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 563.354123] env[63418]: DEBUG oslo.service.loopingcall [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.354123] env[63418]: DEBUG nova.compute.manager [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 563.354123] env[63418]: DEBUG nova.network.neutron [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.401408] env[63418]: DEBUG oslo_concurrency.lockutils [req-f26e1af9-c0cd-4b67-878c-88d52868fef5 req-609d5d30-a3cd-4d3b-8190-b0271661b6e5 service nova] Releasing lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.402840] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquired lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.402840] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.423163] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.711s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.423759] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 563.427022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.223s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.563861] env[63418]: DEBUG nova.network.neutron [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.673106] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 563.934855] env[63418]: DEBUG nova.compute.utils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.936989] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 563.937225] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 563.944728] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.993956] env[63418]: DEBUG nova.policy [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70a99e3be87441259c7a2d6f977aafb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29eb66667c1442fcb136593e6002dd94', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.066753] env[63418]: DEBUG nova.network.neutron [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.129308] env[63418]: INFO nova.scheduler.client.report [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Deleted allocations for instance d255e7af-b50d-4b30-8bec-9e2ee2de2c80 [ 564.174208] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.208097] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.404280] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136b6fbe-7562-45fe-a95c-c16ff4305fea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.414913] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796c2aef-b992-41ce-b76c-a64252cc0e58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.421713] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Successfully created port: 87d9cffe-0dab-4290-91c9-890c7f2d4f32 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.452619] env[63418]: INFO nova.compute.manager [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Rebuilding instance [ 564.456295] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 564.460953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2b2807-e7ac-4268-9c7a-aaedc42db765 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.469860] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb0a6eb-0f4b-4306-b2dc-0ec985270512 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.490023] env[63418]: DEBUG nova.compute.provider_tree [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.511029] env[63418]: DEBUG nova.compute.manager [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 564.512128] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5751e9f4-908e-4a08-83ab-01abc468a090 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.571325] env[63418]: INFO nova.compute.manager [-] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Took 1.22 seconds to deallocate network for instance. [ 564.636335] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fb7e29f9-4097-4861-97cd-6d4f4cc3f7ff tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "d255e7af-b50d-4b30-8bec-9e2ee2de2c80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.492s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.677144] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Releasing lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.677575] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 564.677765] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.678064] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bca6b879-9c86-4e23-a9de-37e6dc504a3a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.688959] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1e78ee-c8aa-4fc8-8b1a-b08f9ddb77d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.720518] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3905f544-be5d-4a2a-a9c6-c7421e41b2f9 could not be found. [ 564.720954] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.720954] env[63418]: INFO nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 564.721136] env[63418]: DEBUG oslo.service.loopingcall [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.721270] env[63418]: DEBUG nova.compute.manager [-] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 564.721403] env[63418]: DEBUG nova.network.neutron [-] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.751240] env[63418]: DEBUG nova.network.neutron [-] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.996783] env[63418]: DEBUG nova.scheduler.client.report [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 565.028606] env[63418]: DEBUG nova.compute.manager [req-238654d4-6d2e-4486-af9b-b56d2efb8393 req-24336955-2685-456b-ba25-600a4ce5e203 service nova] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Received event network-vif-deleted-565d8ffd-fbd2-45c3-918f-51b7f9b07557 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 565.080250] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.139590] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 565.254160] env[63418]: DEBUG nova.network.neutron [-] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.471193] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 565.501731] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.075s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.502666] env[63418]: ERROR nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Traceback (most recent call last): [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self.driver.spawn(context, instance, image_meta, [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] vm_ref = self.build_virtual_machine(instance, [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.502666] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] for vif in network_info: [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return self._sync_wrapper(fn, *args, **kwargs) [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self.wait() [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self[:] = self._gt.wait() [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return self._exit_event.wait() [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] result = hub.switch() [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.503157] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return self.greenlet.switch() [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] result = function(*args, **kwargs) [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] return func(*args, **kwargs) [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] raise e [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] nwinfo = self.network_api.allocate_for_instance( [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] created_port_ids = self._update_ports_for_instance( [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] with excutils.save_and_reraise_exception(): [ 565.503563] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] self.force_reraise() [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] raise self.value [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] updated_port = self._update_port( [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] _ensure_no_port_binding_failure(port) [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] raise exception.PortBindingFailed(port_id=port['id']) [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] nova.exception.PortBindingFailed: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. [ 565.504021] env[63418]: ERROR nova.compute.manager [instance: 781262a7-56ba-466c-86b3-1551905c08bf] [ 565.504348] env[63418]: DEBUG nova.compute.utils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.505441] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.021s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.505676] env[63418]: DEBUG nova.objects.instance [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 565.508703] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Build of instance 781262a7-56ba-466c-86b3-1551905c08bf was re-scheduled: Binding failed for port 902a657c-5430-4645-ab19-a5ec12fe99cf, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.509202] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.509459] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.509635] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquired lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.509819] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.518753] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:37:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='726319986',id=27,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-861448999',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.518753] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.518753] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.518962] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.518962] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.519238] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.519566] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.519736] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.519963] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.520174] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.520376] env[63418]: DEBUG nova.virt.hardware [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.521378] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29606ef7-6d8f-4d1f-bb9f-696d0522499e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.527080] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 565.527497] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d5e072d-21e8-432a-a9d9-2e3b9da1cef3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.534203] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de3f207-863f-47ed-a659-3ffe19dfe94f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.540613] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 565.540613] env[63418]: value = "task-1244696" [ 565.540613] env[63418]: _type = "Task" [ 565.540613] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.558337] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244696, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.581740] env[63418]: ERROR nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 565.581740] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.581740] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.581740] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.581740] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.581740] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.581740] env[63418]: ERROR nova.compute.manager raise self.value [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.581740] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.581740] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.581740] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.582246] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.582246] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.582246] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 565.582246] env[63418]: ERROR nova.compute.manager [ 565.582246] env[63418]: Traceback (most recent call last): [ 565.582246] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.582246] env[63418]: listener.cb(fileno) [ 565.582246] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.582246] env[63418]: result = function(*args, **kwargs) [ 565.582246] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.582246] env[63418]: return func(*args, **kwargs) [ 565.582246] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.582246] env[63418]: raise e [ 565.582246] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.582246] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 565.582246] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.582246] env[63418]: created_port_ids = self._update_ports_for_instance( [ 565.582246] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.582246] env[63418]: with excutils.save_and_reraise_exception(): [ 565.582246] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.582246] env[63418]: self.force_reraise() [ 565.582246] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.582246] env[63418]: raise self.value [ 565.582246] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.582246] env[63418]: updated_port = self._update_port( [ 565.582246] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.582246] env[63418]: _ensure_no_port_binding_failure(port) [ 565.582246] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.582246] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.583123] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 565.583123] env[63418]: Removing descriptor: 15 [ 565.583123] env[63418]: ERROR nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Traceback (most recent call last): [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] yield resources [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self.driver.spawn(context, instance, image_meta, [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.583123] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] vm_ref = self.build_virtual_machine(instance, [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] for vif in network_info: [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return self._sync_wrapper(fn, *args, **kwargs) [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self.wait() [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self[:] = self._gt.wait() [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return self._exit_event.wait() [ 565.583631] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] result = hub.switch() [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return self.greenlet.switch() [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] result = function(*args, **kwargs) [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return func(*args, **kwargs) [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] raise e [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] nwinfo = self.network_api.allocate_for_instance( [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.584211] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] created_port_ids = self._update_ports_for_instance( [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] with excutils.save_and_reraise_exception(): [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self.force_reraise() [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] raise self.value [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] updated_port = self._update_port( [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] _ensure_no_port_binding_failure(port) [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.584587] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] raise exception.PortBindingFailed(port_id=port['id']) [ 565.585575] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 565.585575] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] [ 565.585575] env[63418]: INFO nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Terminating instance [ 565.636182] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquiring lock "8e39c8ef-103a-45f4-8630-bfd51b38445d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.636357] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Lock "8e39c8ef-103a-45f4-8630-bfd51b38445d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.675057] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.762839] env[63418]: INFO nova.compute.manager [-] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Took 1.04 seconds to deallocate network for instance. [ 565.766136] env[63418]: DEBUG nova.compute.claims [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.766136] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.032699] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.051614] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244696, 'name': PowerOffVM_Task, 'duration_secs': 0.22546} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.053895] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 566.054203] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.055282] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc21baf-14e2-4082-b029-dd3b71475e92 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.063103] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 566.063284] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e9ae3b6-244e-4748-b459-630204a81fb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.088626] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquiring lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.088626] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquired lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.088626] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 566.094821] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 566.094821] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 566.094821] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Deleting the datastore file [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 566.094821] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12bc8ce9-0ef0-4b05-9765-a2a40d19a69c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.101156] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 566.101156] env[63418]: value = "task-1244698" [ 566.101156] env[63418]: _type = "Task" [ 566.101156] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.115304] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.143719] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.515019] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bb1033e6-4988-49e3-be64-58544e3137fd tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.516195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.972s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.516394] env[63418]: DEBUG nova.objects.instance [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 566.616120] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147061} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.617574] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.619771] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 566.620095] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 566.620436] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 566.645873] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Releasing lock "refresh_cache-781262a7-56ba-466c-86b3-1551905c08bf" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.646311] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 566.646406] env[63418]: DEBUG nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.646719] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.691583] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.775198] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.194812] env[63418]: DEBUG nova.network.neutron [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.231719] env[63418]: DEBUG nova.compute.manager [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Received event network-changed-87d9cffe-0dab-4290-91c9-890c7f2d4f32 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 567.231719] env[63418]: DEBUG nova.compute.manager [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Refreshing instance network info cache due to event network-changed-87d9cffe-0dab-4290-91c9-890c7f2d4f32. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 567.231719] env[63418]: DEBUG oslo_concurrency.lockutils [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] Acquiring lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.277825] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Releasing lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.278767] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 567.278767] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 567.278992] env[63418]: DEBUG oslo_concurrency.lockutils [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] Acquired lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.279256] env[63418]: DEBUG nova.network.neutron [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Refreshing network info cache for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.280547] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adafc743-d283-4c6c-8734-fee6fe8f2b63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.294792] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65546a3-085d-4619-b715-cb411e162165 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.325175] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 67b5708b-8d02-4816-9455-ea3d9414998a could not be found. [ 567.325426] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 567.325612] env[63418]: INFO nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 567.325869] env[63418]: DEBUG oslo.service.loopingcall [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.327416] env[63418]: DEBUG nova.compute.manager [-] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 567.327523] env[63418]: DEBUG nova.network.neutron [-] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.332441] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "69651fab-08a7-4832-8c91-2fa01d186860" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.332657] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "69651fab-08a7-4832-8c91-2fa01d186860" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.388316] env[63418]: DEBUG nova.network.neutron [-] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.530022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19e5cbad-8c83-4d3f-9308-b10af7e069a5 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.530022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.145s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.531263] env[63418]: INFO nova.compute.claims [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.664371] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.664632] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.664785] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.664954] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.665116] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.665266] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.665469] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.665625] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.665786] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.665945] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.666949] env[63418]: DEBUG nova.virt.hardware [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.668602] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8ec99d-4330-4fc4-b91d-26e10dcc0dbd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.678786] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536bd27b-2ee5-4326-91a8-295a162e9504 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.696238] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 567.702381] env[63418]: DEBUG oslo.service.loopingcall [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.702960] env[63418]: INFO nova.compute.manager [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 781262a7-56ba-466c-86b3-1551905c08bf] Took 1.06 seconds to deallocate network for instance. [ 567.706229] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 567.710055] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0cbe33a6-ec7b-4914-afd6-0d17fa8613f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.731143] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 567.731143] env[63418]: value = "task-1244699" [ 567.731143] env[63418]: _type = "Task" [ 567.731143] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.741291] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244699, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.819091] env[63418]: DEBUG nova.network.neutron [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.891584] env[63418]: DEBUG nova.network.neutron [-] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.037345] env[63418]: DEBUG nova.network.neutron [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.244298] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244699, 'name': CreateVM_Task, 'duration_secs': 0.350874} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.244823] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 568.245417] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.245761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.246223] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 568.248016] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cccbc19d-442e-4642-90ea-9df59023f2f5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.256149] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 568.256149] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a35ffb-1858-5f96-3b8a-eaaaca218f9a" [ 568.256149] env[63418]: _type = "Task" [ 568.256149] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.265237] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a35ffb-1858-5f96-3b8a-eaaaca218f9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.396935] env[63418]: INFO nova.compute.manager [-] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Took 1.07 seconds to deallocate network for instance. [ 568.403014] env[63418]: DEBUG nova.compute.claims [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 568.403014] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.543170] env[63418]: DEBUG oslo_concurrency.lockutils [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] Releasing lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.543458] env[63418]: DEBUG nova.compute.manager [req-39d3c44f-5efd-46b3-a71a-884339507b51 req-9f129c06-c540-4ae8-9c58-79f41fa8b68b service nova] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Received event network-vif-deleted-87d9cffe-0dab-4290-91c9-890c7f2d4f32 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 568.751259] env[63418]: INFO nova.scheduler.client.report [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Deleted allocations for instance 781262a7-56ba-466c-86b3-1551905c08bf [ 568.773391] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a35ffb-1858-5f96-3b8a-eaaaca218f9a, 'name': SearchDatastore_Task, 'duration_secs': 0.010311} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.773749] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.773920] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 568.774327] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.774473] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.774694] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 568.777055] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-098fe23d-ebe2-40e2-ab71-56f9c02d4ce8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.787113] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 568.787113] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 568.787292] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffd62eb6-2755-4b20-a26a-c4be729ed4bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.800772] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 568.800772] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b52f25-7e9c-9281-6bb1-ae0b2e42d026" [ 568.800772] env[63418]: _type = "Task" [ 568.800772] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.824992] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b52f25-7e9c-9281-6bb1-ae0b2e42d026, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.073169] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82af87d1-df55-4922-80cb-5bab303dfc15 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.085987] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d351ae1-64e1-42d6-a7af-0e4fcfc4cff4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.120126] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbac4cab-b0f4-4071-8d48-38f824936550 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.129367] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89685ca5-25ac-4726-b10b-f1e9749b3ce9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.144672] env[63418]: DEBUG nova.compute.provider_tree [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.269688] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2ed9dbca-3bac-4458-9b4a-7204d4690447 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "781262a7-56ba-466c-86b3-1551905c08bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.069s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.316248] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b52f25-7e9c-9281-6bb1-ae0b2e42d026, 'name': SearchDatastore_Task, 'duration_secs': 0.020536} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.318341] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce1cfc38-d22d-481d-84bb-c1773e06b336 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.323604] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 569.323604] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ab1fd5-a509-fbdf-a17f-e519be6e38cc" [ 569.323604] env[63418]: _type = "Task" [ 569.323604] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.332768] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ab1fd5-a509-fbdf-a17f-e519be6e38cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.650550] env[63418]: DEBUG nova.scheduler.client.report [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 569.771503] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 569.836207] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ab1fd5-a509-fbdf-a17f-e519be6e38cc, 'name': SearchDatastore_Task, 'duration_secs': 0.015913} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.836484] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.836724] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 569.836978] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77606ba1-92bb-4733-a5fc-4deda83d3c41 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.846702] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 569.846702] env[63418]: value = "task-1244700" [ 569.846702] env[63418]: _type = "Task" [ 569.846702] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.855733] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.157403] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.628s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.158010] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 570.165457] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.986s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.165719] env[63418]: DEBUG nova.objects.instance [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lazy-loading 'resources' on Instance uuid f185a348-e91d-48d1-970b-473cc253cfdf {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 570.199952] env[63418]: DEBUG nova.compute.manager [None req-c535ee08-e5df-433a-883e-38bd49fe4486 tempest-ServerDiagnosticsV248Test-1168567041 tempest-ServerDiagnosticsV248Test-1168567041-project-admin] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 570.201968] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4348957b-3540-47ad-bd4c-1ada798a7223 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.215309] env[63418]: INFO nova.compute.manager [None req-c535ee08-e5df-433a-883e-38bd49fe4486 tempest-ServerDiagnosticsV248Test-1168567041 tempest-ServerDiagnosticsV248Test-1168567041-project-admin] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Retrieving diagnostics [ 570.216229] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe33a4bb-ecbd-46c4-9703-1dac17eeccee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.294338] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.358531] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244700, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463396} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.358531] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 570.358886] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 570.359240] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a30d3be-859f-4c6a-8440-ea04b60015c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.366419] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 570.366419] env[63418]: value = "task-1244701" [ 570.366419] env[63418]: _type = "Task" [ 570.366419] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.375826] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244701, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.669717] env[63418]: DEBUG nova.compute.utils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.677897] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 570.678193] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 570.680211] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "2edbeef4-be3b-4a1d-8895-8410b71b6b18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.680418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "2edbeef4-be3b-4a1d-8895-8410b71b6b18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.742732] env[63418]: DEBUG nova.policy [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b66bb78486f647e7b181c28fc6c94cca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2a167f3b92f4556825cb4494edd2099', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 570.878567] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244701, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065634} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.881160] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 570.882144] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654afbcf-fa09-46a4-8e9c-4b726807726c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.907274] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 570.910034] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-011c2a79-5c23-48c4-ae83-7a4a744a205c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.933147] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 570.933147] env[63418]: value = "task-1244702" [ 570.933147] env[63418]: _type = "Task" [ 570.933147] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.945259] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244702, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.129017] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Successfully created port: c5dfed01-0c53-473a-9dad-48f1650256e8 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.173248] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e106511a-818c-4888-ba8e-ce2d53da3ba5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.178877] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 571.184736] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4991a0-ab63-4e7a-9b17-51708f573073 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.219437] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70be3256-4099-4218-a7f0-f9f47df5bf2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.228122] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89616564-9957-4991-8981-b0cc8b8fcf74 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.244359] env[63418]: DEBUG nova.compute.provider_tree [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.449478] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244702, 'name': ReconfigVM_Task, 'duration_secs': 0.320593} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.449750] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Reconfigured VM instance instance-0000000d to attach disk [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765/a8c7f192-4672-43cc-8c38-0c33ce633765.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 571.450361] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b08b724-f2d9-4d53-b14d-74453a2e3e20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.459070] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 571.459070] env[63418]: value = "task-1244703" [ 571.459070] env[63418]: _type = "Task" [ 571.459070] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.468795] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244703, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.746990] env[63418]: DEBUG nova.scheduler.client.report [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 571.951298] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.951589] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.951802] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.951976] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.952197] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.955116] env[63418]: INFO nova.compute.manager [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Terminating instance [ 571.974119] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244703, 'name': Rename_Task, 'duration_secs': 0.154478} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.974474] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 571.974739] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4022b1aa-1c36-4f2e-bab2-9fd7b22d5136 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.982674] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 571.982674] env[63418]: value = "task-1244704" [ 571.982674] env[63418]: _type = "Task" [ 571.982674] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.991254] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.015364] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquiring lock "8c78df16-faf1-4007-8492-bdc7cbfc608a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.015591] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Lock "8c78df16-faf1-4007-8492-bdc7cbfc608a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.139184] env[63418]: ERROR nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 572.139184] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.139184] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.139184] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.139184] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.139184] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.139184] env[63418]: ERROR nova.compute.manager raise self.value [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.139184] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.139184] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.139184] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.140836] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.140836] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.140836] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 572.140836] env[63418]: ERROR nova.compute.manager [ 572.140836] env[63418]: Traceback (most recent call last): [ 572.140836] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.140836] env[63418]: listener.cb(fileno) [ 572.140836] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.140836] env[63418]: result = function(*args, **kwargs) [ 572.140836] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.140836] env[63418]: return func(*args, **kwargs) [ 572.140836] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.140836] env[63418]: raise e [ 572.140836] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.140836] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 572.140836] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.140836] env[63418]: created_port_ids = self._update_ports_for_instance( [ 572.140836] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.140836] env[63418]: with excutils.save_and_reraise_exception(): [ 572.140836] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.140836] env[63418]: self.force_reraise() [ 572.140836] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.140836] env[63418]: raise self.value [ 572.140836] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.140836] env[63418]: updated_port = self._update_port( [ 572.140836] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.140836] env[63418]: _ensure_no_port_binding_failure(port) [ 572.140836] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.140836] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.142219] env[63418]: nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 572.142219] env[63418]: Removing descriptor: 15 [ 572.191906] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 572.220544] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.220544] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.220724] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.220922] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.221225] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.221508] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.221816] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.223308] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.223565] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.223914] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.224128] env[63418]: DEBUG nova.virt.hardware [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.224903] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d466cc00-8b2d-4efa-9304-60f7c0fdf164 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.234888] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c14ed2-a25e-4442-b45c-a32c6f1ecdb2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.252492] env[63418]: ERROR nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] Traceback (most recent call last): [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] yield resources [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self.driver.spawn(context, instance, image_meta, [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] vm_ref = self.build_virtual_machine(instance, [ 572.252492] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] for vif in network_info: [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] return self._sync_wrapper(fn, *args, **kwargs) [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self.wait() [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self[:] = self._gt.wait() [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] return self._exit_event.wait() [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 572.253107] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] current.throw(*self._exc) [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] result = function(*args, **kwargs) [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] return func(*args, **kwargs) [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] raise e [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] nwinfo = self.network_api.allocate_for_instance( [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] created_port_ids = self._update_ports_for_instance( [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] with excutils.save_and_reraise_exception(): [ 572.253759] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self.force_reraise() [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] raise self.value [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] updated_port = self._update_port( [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] _ensure_no_port_binding_failure(port) [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] raise exception.PortBindingFailed(port_id=port['id']) [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 572.254522] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] [ 572.254522] env[63418]: INFO nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Terminating instance [ 572.259187] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.094s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.261413] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.077s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.261591] env[63418]: DEBUG nova.objects.instance [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 572.291327] env[63418]: INFO nova.scheduler.client.report [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted allocations for instance f185a348-e91d-48d1-970b-473cc253cfdf [ 572.340457] env[63418]: DEBUG nova.compute.manager [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Received event network-changed-c5dfed01-0c53-473a-9dad-48f1650256e8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 572.340457] env[63418]: DEBUG nova.compute.manager [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Refreshing instance network info cache due to event network-changed-c5dfed01-0c53-473a-9dad-48f1650256e8. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 572.341070] env[63418]: DEBUG oslo_concurrency.lockutils [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] Acquiring lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.341070] env[63418]: DEBUG oslo_concurrency.lockutils [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] Acquired lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.341285] env[63418]: DEBUG nova.network.neutron [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Refreshing network info cache for port c5dfed01-0c53-473a-9dad-48f1650256e8 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 572.468112] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "refresh_cache-ed637299-7e69-43d7-85f9-8fcbd6e90dec" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.468332] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquired lock "refresh_cache-ed637299-7e69-43d7-85f9-8fcbd6e90dec" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.468506] env[63418]: DEBUG nova.network.neutron [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.496602] env[63418]: DEBUG oslo_vmware.api [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244704, 'name': PowerOnVM_Task, 'duration_secs': 0.474889} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.497540] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 572.498104] env[63418]: DEBUG nova.compute.manager [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 572.500103] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6507b573-f669-42dd-a223-afe83edf975b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.757498] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquiring lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.800134] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a0f18217-8d41-4d75-82d0-f776c4c3fe25 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "f185a348-e91d-48d1-970b-473cc253cfdf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.115s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.871297] env[63418]: DEBUG nova.network.neutron [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.935190] env[63418]: DEBUG nova.network.neutron [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.011357] env[63418]: DEBUG nova.network.neutron [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.025218] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.099095] env[63418]: DEBUG nova.network.neutron [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.272425] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0aa0dadc-4e58-418c-a5fe-7720887aa5aa tempest-ServersAdmin275Test-1349321556 tempest-ServersAdmin275Test-1349321556-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.274729] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.874s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.276820] env[63418]: INFO nova.compute.claims [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.438863] env[63418]: DEBUG oslo_concurrency.lockutils [req-05024e73-a408-4740-9368-9d86dd485048 req-fbf0614e-e4a4-4874-ad18-78fe6311a9c1 service nova] Releasing lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.439324] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquired lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.439510] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.603136] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Releasing lock "refresh_cache-ed637299-7e69-43d7-85f9-8fcbd6e90dec" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.605226] env[63418]: DEBUG nova.compute.manager [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 573.605473] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.606600] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2a9531-06f9-4e44-9d18-5a540b37657f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.616282] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 573.616551] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-338d5ec7-293d-46cd-ae05-1b9657684a49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.625083] env[63418]: DEBUG oslo_vmware.api [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 573.625083] env[63418]: value = "task-1244705" [ 573.625083] env[63418]: _type = "Task" [ 573.625083] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.641241] env[63418]: DEBUG oslo_vmware.api [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244705, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.713143] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "6a80e0c1-2088-4de3-9032-d05ff48c3fcb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.713636] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "6a80e0c1-2088-4de3-9032-d05ff48c3fcb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.744167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "a8c7f192-4672-43cc-8c38-0c33ce633765" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.744434] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "a8c7f192-4672-43cc-8c38-0c33ce633765" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.744644] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "a8c7f192-4672-43cc-8c38-0c33ce633765-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.744817] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "a8c7f192-4672-43cc-8c38-0c33ce633765-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.744971] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "a8c7f192-4672-43cc-8c38-0c33ce633765-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.747858] env[63418]: INFO nova.compute.manager [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Terminating instance [ 573.965917] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.045564] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.135670] env[63418]: DEBUG oslo_vmware.api [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244705, 'name': PowerOffVM_Task, 'duration_secs': 0.143952} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.135928] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 574.136112] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 574.136476] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f51480f1-7a5e-4315-ad7f-17c76c062304 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.166615] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 574.166615] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 574.166750] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Deleting the datastore file [datastore2] ed637299-7e69-43d7-85f9-8fcbd6e90dec {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 574.166984] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2618a46f-a746-4472-a30b-c1b086dd418b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.175835] env[63418]: DEBUG oslo_vmware.api [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for the task: (returnval){ [ 574.175835] env[63418]: value = "task-1244707" [ 574.175835] env[63418]: _type = "Task" [ 574.175835] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.185647] env[63418]: DEBUG oslo_vmware.api [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.214035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "7e43d259-f361-43d8-8f03-72b303680478" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.214035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "7e43d259-f361-43d8-8f03-72b303680478" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.214035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "7e43d259-f361-43d8-8f03-72b303680478-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.214035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "7e43d259-f361-43d8-8f03-72b303680478-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.214294] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "7e43d259-f361-43d8-8f03-72b303680478-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.216734] env[63418]: INFO nova.compute.manager [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Terminating instance [ 574.256667] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "refresh_cache-a8c7f192-4672-43cc-8c38-0c33ce633765" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.256868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquired lock "refresh_cache-a8c7f192-4672-43cc-8c38-0c33ce633765" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.257059] env[63418]: DEBUG nova.network.neutron [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.419952] env[63418]: DEBUG nova.compute.manager [req-61e2ed67-f129-48bd-b9e2-1228271a0e29 req-1d24c688-29fe-4a3d-be13-7f3a30b14f90 service nova] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Received event network-vif-deleted-c5dfed01-0c53-473a-9dad-48f1650256e8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 574.551552] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Releasing lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.551976] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 574.552186] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.552472] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f224da6e-e840-4dd7-8394-9bed94c3bd4c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.564591] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fccecf-ebee-4891-9c4b-ca37a8678472 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.591731] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 58850986-3ebf-465f-b832-9b245742a34a could not be found. [ 574.592312] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.592545] env[63418]: INFO nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 574.592834] env[63418]: DEBUG oslo.service.loopingcall [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.595711] env[63418]: DEBUG nova.compute.manager [-] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 574.595789] env[63418]: DEBUG nova.network.neutron [-] [instance: 58850986-3ebf-465f-b832-9b245742a34a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.612041] env[63418]: DEBUG nova.network.neutron [-] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.692642] env[63418]: DEBUG oslo_vmware.api [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Task: {'id': task-1244707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094264} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.695334] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 574.695529] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 574.695703] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.695872] env[63418]: INFO nova.compute.manager [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Took 1.09 seconds to destroy the instance on the hypervisor. [ 574.696122] env[63418]: DEBUG oslo.service.loopingcall [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.696495] env[63418]: DEBUG nova.compute.manager [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 574.696591] env[63418]: DEBUG nova.network.neutron [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.713956] env[63418]: DEBUG nova.network.neutron [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.719955] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1926e147-875d-44ad-9047-290502a94ce3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.723548] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "refresh_cache-7e43d259-f361-43d8-8f03-72b303680478" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.723759] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquired lock "refresh_cache-7e43d259-f361-43d8-8f03-72b303680478" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.723960] env[63418]: DEBUG nova.network.neutron [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.730962] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf883ab9-ea54-4d98-a865-7d2bd819a876 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.764106] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fb0f52-dc0b-481e-ab08-1352b6bd8332 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.774579] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f870325-080b-4d4c-88a9-9a28c5960fed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.791859] env[63418]: DEBUG nova.compute.provider_tree [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.791859] env[63418]: DEBUG nova.network.neutron [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.852267] env[63418]: DEBUG nova.network.neutron [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.114880] env[63418]: DEBUG nova.network.neutron [-] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.217476] env[63418]: DEBUG nova.network.neutron [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.246025] env[63418]: DEBUG nova.network.neutron [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.294843] env[63418]: DEBUG nova.scheduler.client.report [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 575.326673] env[63418]: DEBUG nova.network.neutron [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.357024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Releasing lock "refresh_cache-a8c7f192-4672-43cc-8c38-0c33ce633765" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.357024] env[63418]: DEBUG nova.compute.manager [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 575.357024] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.357024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3ea701-cf85-402f-b757-ada62865df0d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.367118] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 575.367373] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36352ec4-0ca0-4afc-948a-2746543ce75d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.374548] env[63418]: DEBUG oslo_vmware.api [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 575.374548] env[63418]: value = "task-1244708" [ 575.374548] env[63418]: _type = "Task" [ 575.374548] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.386313] env[63418]: DEBUG oslo_vmware.api [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.559466] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.559713] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.618367] env[63418]: INFO nova.compute.manager [-] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Took 1.02 seconds to deallocate network for instance. [ 575.624037] env[63418]: DEBUG nova.compute.claims [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.627110] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.721803] env[63418]: INFO nova.compute.manager [-] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Took 1.02 seconds to deallocate network for instance. [ 575.802599] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.803182] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 575.808471] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.743s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.808471] env[63418]: INFO nova.compute.claims [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.829796] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Releasing lock "refresh_cache-7e43d259-f361-43d8-8f03-72b303680478" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.833025] env[63418]: DEBUG nova.compute.manager [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 575.833025] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.833025] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c84667-ddb8-458c-a640-dd0ed60f1e4c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.841919] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 575.842199] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6984a1f9-afb8-40e0-9279-db2dcecd05c0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.854225] env[63418]: DEBUG oslo_vmware.api [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 575.854225] env[63418]: value = "task-1244709" [ 575.854225] env[63418]: _type = "Task" [ 575.854225] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.865654] env[63418]: DEBUG oslo_vmware.api [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.885130] env[63418]: DEBUG oslo_vmware.api [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244708, 'name': PowerOffVM_Task, 'duration_secs': 0.225954} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.885464] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 575.885665] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 575.885969] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5cc29ad1-bce9-4a07-9bf0-bf7adfaa5566 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.916995] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 575.917303] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 575.917449] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Deleting the datastore file [datastore2] a8c7f192-4672-43cc-8c38-0c33ce633765 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.917738] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d74ec57-f10d-4b8a-bf54-3761cfabb6a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.928025] env[63418]: DEBUG oslo_vmware.api [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for the task: (returnval){ [ 575.928025] env[63418]: value = "task-1244711" [ 575.928025] env[63418]: _type = "Task" [ 575.928025] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.935055] env[63418]: DEBUG oslo_vmware.api [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.232485] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.314179] env[63418]: DEBUG nova.compute.utils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 576.315808] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 576.315935] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 576.367963] env[63418]: DEBUG oslo_vmware.api [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244709, 'name': PowerOffVM_Task, 'duration_secs': 0.140856} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.367963] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 576.367963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 576.367963] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e5ed9c9-384e-450e-b4fc-614a3e2e61a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.397417] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 576.397941] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 576.397941] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleting the datastore file [datastore1] 7e43d259-f361-43d8-8f03-72b303680478 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 576.398618] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4b12bb5-aa46-495c-a285-e73421bd1fd3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.405647] env[63418]: DEBUG oslo_vmware.api [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for the task: (returnval){ [ 576.405647] env[63418]: value = "task-1244713" [ 576.405647] env[63418]: _type = "Task" [ 576.405647] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.411505] env[63418]: DEBUG nova.policy [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ff64612530b451fb41100a5aa601be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d52f1fceb24234a8d967038b43c857', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.416359] env[63418]: DEBUG oslo_vmware.api [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.435848] env[63418]: DEBUG oslo_vmware.api [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Task: {'id': task-1244711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114349} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.436124] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 576.436321] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 576.436514] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.436691] env[63418]: INFO nova.compute.manager [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Took 1.08 seconds to destroy the instance on the hypervisor. [ 576.436940] env[63418]: DEBUG oslo.service.loopingcall [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.437159] env[63418]: DEBUG nova.compute.manager [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 576.437262] env[63418]: DEBUG nova.network.neutron [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.456704] env[63418]: DEBUG nova.network.neutron [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.700910] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Successfully created port: d735e501-e398-41b4-a4f6-39af715bd5c3 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.823613] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 576.920805] env[63418]: DEBUG oslo_vmware.api [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Task: {'id': task-1244713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100505} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.923732] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 576.924099] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 576.926028] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.926028] env[63418]: INFO nova.compute.manager [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Took 1.09 seconds to destroy the instance on the hypervisor. [ 576.926028] env[63418]: DEBUG oslo.service.loopingcall [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.926028] env[63418]: DEBUG nova.compute.manager [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 576.926028] env[63418]: DEBUG nova.network.neutron [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.951428] env[63418]: DEBUG nova.network.neutron [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.961075] env[63418]: DEBUG nova.network.neutron [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.336781] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3309935d-8d92-45c3-b32d-030555b7542a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.349447] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afd90e6-7566-45d6-a7cc-87f5421bd0ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.382349] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e25cda3-9293-4712-a1a9-1a2de6c9f09f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.391372] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c616de46-3f7f-47b9-b8dd-a98ed9fe4ce7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.406280] env[63418]: DEBUG nova.compute.provider_tree [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.455355] env[63418]: DEBUG nova.network.neutron [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.464793] env[63418]: INFO nova.compute.manager [-] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Took 1.03 seconds to deallocate network for instance. [ 577.493779] env[63418]: DEBUG nova.compute.manager [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Received event network-changed-d735e501-e398-41b4-a4f6-39af715bd5c3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 577.494160] env[63418]: DEBUG nova.compute.manager [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Refreshing instance network info cache due to event network-changed-d735e501-e398-41b4-a4f6-39af715bd5c3. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 577.494310] env[63418]: DEBUG oslo_concurrency.lockutils [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] Acquiring lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.494408] env[63418]: DEBUG oslo_concurrency.lockutils [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] Acquired lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.494571] env[63418]: DEBUG nova.network.neutron [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Refreshing network info cache for port d735e501-e398-41b4-a4f6-39af715bd5c3 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 577.789409] env[63418]: ERROR nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 577.789409] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 577.789409] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.789409] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.789409] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.789409] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.789409] env[63418]: ERROR nova.compute.manager raise self.value [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.789409] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.789409] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.789409] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.789960] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.789960] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.789960] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 577.789960] env[63418]: ERROR nova.compute.manager [ 577.789960] env[63418]: Traceback (most recent call last): [ 577.789960] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.789960] env[63418]: listener.cb(fileno) [ 577.789960] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.789960] env[63418]: result = function(*args, **kwargs) [ 577.789960] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.789960] env[63418]: return func(*args, **kwargs) [ 577.789960] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 577.789960] env[63418]: raise e [ 577.789960] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 577.789960] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 577.789960] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.789960] env[63418]: created_port_ids = self._update_ports_for_instance( [ 577.789960] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.789960] env[63418]: with excutils.save_and_reraise_exception(): [ 577.789960] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.789960] env[63418]: self.force_reraise() [ 577.789960] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.789960] env[63418]: raise self.value [ 577.789960] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.789960] env[63418]: updated_port = self._update_port( [ 577.789960] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.789960] env[63418]: _ensure_no_port_binding_failure(port) [ 577.789960] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.789960] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.790849] env[63418]: nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 577.790849] env[63418]: Removing descriptor: 15 [ 577.838609] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 577.867747] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.868053] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.868244] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.868456] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.868630] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.868818] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.869177] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.873102] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.873102] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.873102] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.873102] env[63418]: DEBUG nova.virt.hardware [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.873102] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afdb7b0-7861-4505-9935-291bd429df51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.879585] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0ec91f-d975-4dcd-926d-9dd9e4f61caa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.896039] env[63418]: ERROR nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Traceback (most recent call last): [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] yield resources [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self.driver.spawn(context, instance, image_meta, [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] vm_ref = self.build_virtual_machine(instance, [ 577.896039] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] for vif in network_info: [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] return self._sync_wrapper(fn, *args, **kwargs) [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self.wait() [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self[:] = self._gt.wait() [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] return self._exit_event.wait() [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 577.896577] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] current.throw(*self._exc) [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] result = function(*args, **kwargs) [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] return func(*args, **kwargs) [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] raise e [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] nwinfo = self.network_api.allocate_for_instance( [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] created_port_ids = self._update_ports_for_instance( [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] with excutils.save_and_reraise_exception(): [ 577.897157] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self.force_reraise() [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] raise self.value [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] updated_port = self._update_port( [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] _ensure_no_port_binding_failure(port) [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] raise exception.PortBindingFailed(port_id=port['id']) [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 577.897787] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] [ 577.897787] env[63418]: INFO nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Terminating instance [ 577.908943] env[63418]: DEBUG nova.scheduler.client.report [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 577.959886] env[63418]: INFO nova.compute.manager [-] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Took 1.03 seconds to deallocate network for instance. [ 577.974325] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.013064] env[63418]: DEBUG nova.network.neutron [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.104439] env[63418]: DEBUG nova.network.neutron [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.536259] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.536502] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.730s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.537486] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 578.544136] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.547839] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.434s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.550573] env[63418]: INFO nova.compute.claims [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.609506] env[63418]: DEBUG oslo_concurrency.lockutils [req-0b769635-96c0-4880-87d1-5e3c9069edf8 req-6fa6ba29-7bfe-4b0c-8690-90f3ef860134 service nova] Releasing lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.609911] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.610114] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.717047] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquiring lock "c79aa3d1-bbca-45cf-9f29-40a2720c4437" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.717293] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Lock "c79aa3d1-bbca-45cf-9f29-40a2720c4437" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.049972] env[63418]: DEBUG nova.compute.utils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.051143] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 579.051318] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 579.113158] env[63418]: DEBUG nova.policy [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42898bae47f14a35bb506d0d24f4ef7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5917fe6892094b1db75d1dae77a3c154', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.135653] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.232017] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.379262] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Successfully created port: ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.518627] env[63418]: DEBUG nova.compute.manager [req-aa72c20c-b59b-43a8-8292-586f5d199d85 req-ed3b830c-5ab5-41f3-865a-10f5fbe3568f service nova] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Received event network-vif-deleted-d735e501-e398-41b4-a4f6-39af715bd5c3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 579.554218] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 579.737045] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.737045] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 579.737045] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.737902] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2250feeb-fedd-49ac-8df1-c8cf06861fd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.747936] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e6baab-8d0d-4b82-b8ea-a81277dd1961 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.776514] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34cf7345-c244-44bd-aa1f-4788ae8f6da5 could not be found. [ 579.776896] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.777168] env[63418]: INFO nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 579.777477] env[63418]: DEBUG oslo.service.loopingcall [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.778117] env[63418]: DEBUG nova.compute.manager [-] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 579.778294] env[63418]: DEBUG nova.network.neutron [-] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.936631] env[63418]: DEBUG nova.network.neutron [-] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.033936] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39354ba-328a-45e9-a317-c53a43cb395f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.045298] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d38f49e-ac92-457b-9b15-fb953a71bea2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.079584] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46c9353-6dc7-46d7-b011-4e477334e925 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.088581] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf925c16-f839-4ffb-bc08-ff0e8710342a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.104189] env[63418]: DEBUG nova.compute.provider_tree [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.246770] env[63418]: ERROR nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 580.246770] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 580.246770] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.246770] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.246770] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.246770] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.246770] env[63418]: ERROR nova.compute.manager raise self.value [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.246770] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.246770] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.246770] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.247399] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.247399] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.247399] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 580.247399] env[63418]: ERROR nova.compute.manager [ 580.247399] env[63418]: Traceback (most recent call last): [ 580.247399] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.247399] env[63418]: listener.cb(fileno) [ 580.247399] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.247399] env[63418]: result = function(*args, **kwargs) [ 580.247399] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.247399] env[63418]: return func(*args, **kwargs) [ 580.247399] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 580.247399] env[63418]: raise e [ 580.247399] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 580.247399] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 580.247399] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.247399] env[63418]: created_port_ids = self._update_ports_for_instance( [ 580.247399] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.247399] env[63418]: with excutils.save_and_reraise_exception(): [ 580.247399] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.247399] env[63418]: self.force_reraise() [ 580.247399] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.247399] env[63418]: raise self.value [ 580.247399] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.247399] env[63418]: updated_port = self._update_port( [ 580.247399] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.247399] env[63418]: _ensure_no_port_binding_failure(port) [ 580.247399] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.247399] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.248380] env[63418]: nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 580.248380] env[63418]: Removing descriptor: 15 [ 580.438479] env[63418]: DEBUG nova.network.neutron [-] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.581787] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 580.609928] env[63418]: DEBUG nova.scheduler.client.report [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 580.623060] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.623216] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.623465] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.623710] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.623916] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.624842] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.624842] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.624842] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.624842] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.625047] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.625197] env[63418]: DEBUG nova.virt.hardware [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.626132] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b12fdcd-9d57-42b8-8983-38803f049817 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.636717] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b23c9ab-803e-4693-8499-87873ca32981 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.653114] env[63418]: ERROR nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Traceback (most recent call last): [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] yield resources [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self.driver.spawn(context, instance, image_meta, [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] vm_ref = self.build_virtual_machine(instance, [ 580.653114] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] for vif in network_info: [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] return self._sync_wrapper(fn, *args, **kwargs) [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self.wait() [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self[:] = self._gt.wait() [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] return self._exit_event.wait() [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 580.653570] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] current.throw(*self._exc) [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] result = function(*args, **kwargs) [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] return func(*args, **kwargs) [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] raise e [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] nwinfo = self.network_api.allocate_for_instance( [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] created_port_ids = self._update_ports_for_instance( [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] with excutils.save_and_reraise_exception(): [ 580.653964] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self.force_reraise() [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] raise self.value [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] updated_port = self._update_port( [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] _ensure_no_port_binding_failure(port) [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] raise exception.PortBindingFailed(port_id=port['id']) [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 580.654371] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] [ 580.654371] env[63418]: INFO nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Terminating instance [ 580.942365] env[63418]: INFO nova.compute.manager [-] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Took 1.16 seconds to deallocate network for instance. [ 580.945203] env[63418]: DEBUG nova.compute.claims [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.945463] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.114912] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.115214] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 581.118110] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.312s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.118346] env[63418]: DEBUG nova.objects.instance [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lazy-loading 'resources' on Instance uuid ba67658a-668e-4fca-aefe-e838f7b05e2a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 581.156982] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquiring lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.157183] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquired lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.157356] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.587102] env[63418]: DEBUG nova.compute.manager [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Received event network-changed-ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 581.587102] env[63418]: DEBUG nova.compute.manager [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Refreshing instance network info cache due to event network-changed-ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 581.587102] env[63418]: DEBUG oslo_concurrency.lockutils [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] Acquiring lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.621621] env[63418]: DEBUG nova.compute.utils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.626178] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 581.626366] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.669644] env[63418]: DEBUG nova.policy [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aba989cf983640768417434abd994fba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76213dff9a794429a79bc8dce7f00fa7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.678451] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.804373] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.945170] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Successfully created port: f9ebb350-3730-4f9d-b221-b6af2642a7b2 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.060491] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bca184-51e0-40a3-af2d-8aec4b394b16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.067695] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90030e7-a0d4-480c-a031-a3a47bd27f63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.100987] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41e736a-f6d6-4fcb-88af-be2ea15067bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.109562] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff84d75-1ae5-4439-9071-3161a593820c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.123753] env[63418]: DEBUG nova.compute.provider_tree [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.126797] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 582.307273] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Releasing lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.307736] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 582.307925] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.308256] env[63418]: DEBUG oslo_concurrency.lockutils [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] Acquired lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.308427] env[63418]: DEBUG nova.network.neutron [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Refreshing network info cache for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 582.309504] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e838fda6-0a62-42ad-afac-4017b90949ca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.325354] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910886c2-bc0d-44f6-bc20-f1fef0652f3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.348836] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5caab97d-833e-4356-af51-4a991d0e5b44 could not be found. [ 582.349119] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.349248] env[63418]: INFO nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Took 0.04 seconds to destroy the instance on the hypervisor. [ 582.349483] env[63418]: DEBUG oslo.service.loopingcall [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.349692] env[63418]: DEBUG nova.compute.manager [-] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 582.349782] env[63418]: DEBUG nova.network.neutron [-] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.371603] env[63418]: DEBUG nova.network.neutron [-] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.626999] env[63418]: DEBUG nova.scheduler.client.report [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 582.811182] env[63418]: ERROR nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 582.811182] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.811182] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.811182] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.811182] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.811182] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.811182] env[63418]: ERROR nova.compute.manager raise self.value [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.811182] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.811182] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.811182] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.811688] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.811688] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.811688] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 582.811688] env[63418]: ERROR nova.compute.manager [ 582.811688] env[63418]: Traceback (most recent call last): [ 582.811688] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.811688] env[63418]: listener.cb(fileno) [ 582.811688] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.811688] env[63418]: result = function(*args, **kwargs) [ 582.811688] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.811688] env[63418]: return func(*args, **kwargs) [ 582.811688] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.811688] env[63418]: raise e [ 582.811688] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.811688] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 582.811688] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.811688] env[63418]: created_port_ids = self._update_ports_for_instance( [ 582.811688] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.811688] env[63418]: with excutils.save_and_reraise_exception(): [ 582.811688] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.811688] env[63418]: self.force_reraise() [ 582.811688] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.811688] env[63418]: raise self.value [ 582.811688] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.811688] env[63418]: updated_port = self._update_port( [ 582.811688] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.811688] env[63418]: _ensure_no_port_binding_failure(port) [ 582.811688] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.811688] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.812634] env[63418]: nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 582.812634] env[63418]: Removing descriptor: 15 [ 582.828302] env[63418]: DEBUG nova.network.neutron [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.874628] env[63418]: DEBUG nova.network.neutron [-] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.915617] env[63418]: DEBUG nova.network.neutron [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.136250] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.018s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.139408] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.197s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.144735] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 583.161712] env[63418]: INFO nova.scheduler.client.report [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Deleted allocations for instance ba67658a-668e-4fca-aefe-e838f7b05e2a [ 583.172029] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.172265] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.172412] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.172584] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.172720] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.172858] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.173104] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.173275] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.174103] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.174103] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.174103] env[63418]: DEBUG nova.virt.hardware [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.174645] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b78080-c27c-4e00-a070-2cde9bd30f51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.184701] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe93b7f-c461-4725-9cec-47a4816b9859 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.199331] env[63418]: ERROR nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Traceback (most recent call last): [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] yield resources [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self.driver.spawn(context, instance, image_meta, [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] vm_ref = self.build_virtual_machine(instance, [ 583.199331] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] for vif in network_info: [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] return self._sync_wrapper(fn, *args, **kwargs) [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self.wait() [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self[:] = self._gt.wait() [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] return self._exit_event.wait() [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 583.199801] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] current.throw(*self._exc) [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] result = function(*args, **kwargs) [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] return func(*args, **kwargs) [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] raise e [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] nwinfo = self.network_api.allocate_for_instance( [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] created_port_ids = self._update_ports_for_instance( [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] with excutils.save_and_reraise_exception(): [ 583.200235] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self.force_reraise() [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] raise self.value [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] updated_port = self._update_port( [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] _ensure_no_port_binding_failure(port) [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] raise exception.PortBindingFailed(port_id=port['id']) [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 583.200697] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] [ 583.200697] env[63418]: INFO nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Terminating instance [ 583.376581] env[63418]: INFO nova.compute.manager [-] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Took 1.03 seconds to deallocate network for instance. [ 583.378919] env[63418]: DEBUG nova.compute.claims [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.379120] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.419034] env[63418]: DEBUG oslo_concurrency.lockutils [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] Releasing lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.419223] env[63418]: DEBUG nova.compute.manager [req-461e77bf-58fe-4357-a0cc-d06d5c2b0eee req-244f1a08-652d-451e-8478-5d00394b42cd service nova] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Received event network-vif-deleted-ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 583.643299] env[63418]: DEBUG nova.compute.manager [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Received event network-changed-f9ebb350-3730-4f9d-b221-b6af2642a7b2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 583.643608] env[63418]: DEBUG nova.compute.manager [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Refreshing instance network info cache due to event network-changed-f9ebb350-3730-4f9d-b221-b6af2642a7b2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 583.643899] env[63418]: DEBUG oslo_concurrency.lockutils [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] Acquiring lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.644383] env[63418]: DEBUG oslo_concurrency.lockutils [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] Acquired lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.644565] env[63418]: DEBUG nova.network.neutron [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Refreshing network info cache for port f9ebb350-3730-4f9d-b221-b6af2642a7b2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.674873] env[63418]: DEBUG oslo_concurrency.lockutils [None req-69758c85-da62-4053-bfc0-63f19f34f4f0 tempest-ServersAdmin275Test-98490622 tempest-ServersAdmin275Test-98490622-project-member] Lock "ba67658a-668e-4fca-aefe-e838f7b05e2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.246s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.703114] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquiring lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.036580] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091f8471-dcc3-4576-b639-622a2a6d2f61 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.044440] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c94494-b421-4489-92ca-dd3a858805b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.076184] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5857ee51-b171-4464-a9e4-bd669499ee23 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.083779] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f62ace-ed97-4845-956f-e060e672b7e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.096868] env[63418]: DEBUG nova.compute.provider_tree [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.163652] env[63418]: DEBUG nova.network.neutron [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.242620] env[63418]: DEBUG nova.network.neutron [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.602589] env[63418]: DEBUG nova.scheduler.client.report [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 584.747547] env[63418]: DEBUG oslo_concurrency.lockutils [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] Releasing lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.747547] env[63418]: DEBUG nova.compute.manager [req-8b8a8a4c-4028-42f9-8710-6ff3023d16ca req-2b5e400a-60c1-445c-954c-9fa52404c481 service nova] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Received event network-vif-deleted-f9ebb350-3730-4f9d-b221-b6af2642a7b2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 584.748422] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquired lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.748422] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.105645] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.106338] env[63418]: ERROR nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] Traceback (most recent call last): [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self.driver.spawn(context, instance, image_meta, [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] vm_ref = self.build_virtual_machine(instance, [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.106338] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] for vif in network_info: [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return self._sync_wrapper(fn, *args, **kwargs) [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self.wait() [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self[:] = self._gt.wait() [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return self._exit_event.wait() [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] result = hub.switch() [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.106735] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return self.greenlet.switch() [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] result = function(*args, **kwargs) [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] return func(*args, **kwargs) [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] raise e [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] nwinfo = self.network_api.allocate_for_instance( [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] created_port_ids = self._update_ports_for_instance( [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] with excutils.save_and_reraise_exception(): [ 585.107160] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] self.force_reraise() [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] raise self.value [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] updated_port = self._update_port( [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] _ensure_no_port_binding_failure(port) [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] raise exception.PortBindingFailed(port_id=port['id']) [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] nova.exception.PortBindingFailed: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. [ 585.107589] env[63418]: ERROR nova.compute.manager [instance: f714b41e-b537-4feb-bd71-53b5db089485] [ 585.107964] env[63418]: DEBUG nova.compute.utils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.108750] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.904s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.110419] env[63418]: INFO nova.compute.claims [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.112720] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Build of instance f714b41e-b537-4feb-bd71-53b5db089485 was re-scheduled: Binding failed for port 2163b943-c087-4595-af3d-2f900fff8221, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 585.113308] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 585.113523] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquiring lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.113700] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Acquired lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.113933] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.272988] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.368574] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.635373] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.711076] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.873043] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Releasing lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.873512] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 585.873706] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.874027] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed8d3d87-15e4-4a32-b174-d21c8988240c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.886478] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e3cefa-972f-4cd0-83f2-6b14ee69b9d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.911296] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1b24de1-3d7f-40b7-9183-c4b1f35f8c94 could not be found. [ 585.911525] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 585.911706] env[63418]: INFO nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Took 0.04 seconds to destroy the instance on the hypervisor. [ 585.911947] env[63418]: DEBUG oslo.service.loopingcall [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.912182] env[63418]: DEBUG nova.compute.manager [-] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 585.912274] env[63418]: DEBUG nova.network.neutron [-] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.927288] env[63418]: DEBUG nova.network.neutron [-] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.213641] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Releasing lock "refresh_cache-f714b41e-b537-4feb-bd71-53b5db089485" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.213871] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 586.214063] env[63418]: DEBUG nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.214220] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.232305] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.429446] env[63418]: DEBUG nova.network.neutron [-] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.592906] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8234f494-b77c-4fc5-9b33-2823042523d6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.603153] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c083bc-31e2-4334-912c-541510cb9ec6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.641854] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415e4ff8-2b43-4b8c-bc40-339df003677c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.650852] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e83eb28-f885-40f9-9f1d-9ade193adc35 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.672018] env[63418]: DEBUG nova.compute.provider_tree [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.739753] env[63418]: DEBUG nova.network.neutron [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.932218] env[63418]: INFO nova.compute.manager [-] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Took 1.02 seconds to deallocate network for instance. [ 586.934554] env[63418]: DEBUG nova.compute.claims [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 586.934739] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.173153] env[63418]: DEBUG nova.scheduler.client.report [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 587.244736] env[63418]: INFO nova.compute.manager [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] [instance: f714b41e-b537-4feb-bd71-53b5db089485] Took 1.03 seconds to deallocate network for instance. [ 587.678115] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.678703] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 587.682143] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.602s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.682389] env[63418]: DEBUG nova.objects.instance [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lazy-loading 'resources' on Instance uuid 01deeb21-0e27-497f-9b85-c85949a3533d {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 588.381716] env[63418]: DEBUG nova.compute.utils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.386491] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 588.386684] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.492986] env[63418]: DEBUG nova.policy [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f21cde13021c42589936d984cee5f8cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '048efdbecdec42f596db63da0b2f0c2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.869990] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Successfully created port: e2d74ac4-1808-4ebe-ac7e-3c64149c766c {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.889087] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 588.926342] env[63418]: INFO nova.scheduler.client.report [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Deleted allocations for instance f714b41e-b537-4feb-bd71-53b5db089485 [ 588.943097] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2769cab9-9d58-41fa-a24b-c7d909bb2330 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.952812] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efdfbc5a-8032-4f5a-8f17-8e9035fed7a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.989592] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a166bc-9dd0-4d6e-b035-ad3eda3d114c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.998758] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5542ad9c-35d2-4d74-bfe5-ca0594405295 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.015289] env[63418]: DEBUG nova.compute.provider_tree [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.438180] env[63418]: DEBUG oslo_concurrency.lockutils [None req-61d2a052-1db6-45e9-9faf-d87e24c5d50b tempest-VolumesAssistedSnapshotsTest-1173310601 tempest-VolumesAssistedSnapshotsTest-1173310601-project-member] Lock "f714b41e-b537-4feb-bd71-53b5db089485" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.465s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.521481] env[63418]: DEBUG nova.scheduler.client.report [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 589.699329] env[63418]: DEBUG nova.compute.manager [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Received event network-changed-e2d74ac4-1808-4ebe-ac7e-3c64149c766c {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 589.699513] env[63418]: DEBUG nova.compute.manager [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Refreshing instance network info cache due to event network-changed-e2d74ac4-1808-4ebe-ac7e-3c64149c766c. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 589.699747] env[63418]: DEBUG oslo_concurrency.lockutils [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] Acquiring lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.699887] env[63418]: DEBUG oslo_concurrency.lockutils [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] Acquired lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.700750] env[63418]: DEBUG nova.network.neutron [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Refreshing network info cache for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 589.789247] env[63418]: ERROR nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 589.789247] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.789247] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.789247] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.789247] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.789247] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.789247] env[63418]: ERROR nova.compute.manager raise self.value [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.789247] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.789247] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.789247] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.789673] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.789673] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.789673] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 589.789673] env[63418]: ERROR nova.compute.manager [ 589.789673] env[63418]: Traceback (most recent call last): [ 589.789673] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.789673] env[63418]: listener.cb(fileno) [ 589.789673] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.789673] env[63418]: result = function(*args, **kwargs) [ 589.789673] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.789673] env[63418]: return func(*args, **kwargs) [ 589.789673] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.789673] env[63418]: raise e [ 589.789673] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.789673] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 589.789673] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.789673] env[63418]: created_port_ids = self._update_ports_for_instance( [ 589.789673] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.789673] env[63418]: with excutils.save_and_reraise_exception(): [ 589.789673] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.789673] env[63418]: self.force_reraise() [ 589.789673] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.789673] env[63418]: raise self.value [ 589.789673] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.789673] env[63418]: updated_port = self._update_port( [ 589.789673] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.789673] env[63418]: _ensure_no_port_binding_failure(port) [ 589.789673] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.789673] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.790311] env[63418]: nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 589.790311] env[63418]: Removing descriptor: 15 [ 589.910989] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 589.940320] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 589.945213] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.945482] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.945668] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.946391] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.946391] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.946391] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.946391] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.946608] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.946769] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.946955] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.947175] env[63418]: DEBUG nova.virt.hardware [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.948179] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9f7264-ae03-4330-adf9-df859499185a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.957760] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da88533-23f7-4274-b82f-4e499e81c21b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.972781] env[63418]: ERROR nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Traceback (most recent call last): [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] yield resources [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self.driver.spawn(context, instance, image_meta, [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] vm_ref = self.build_virtual_machine(instance, [ 589.972781] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] for vif in network_info: [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] return self._sync_wrapper(fn, *args, **kwargs) [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self.wait() [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self[:] = self._gt.wait() [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] return self._exit_event.wait() [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 589.973172] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] current.throw(*self._exc) [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] result = function(*args, **kwargs) [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] return func(*args, **kwargs) [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] raise e [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] nwinfo = self.network_api.allocate_for_instance( [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] created_port_ids = self._update_ports_for_instance( [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] with excutils.save_and_reraise_exception(): [ 589.973425] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self.force_reraise() [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] raise self.value [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] updated_port = self._update_port( [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] _ensure_no_port_binding_failure(port) [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] raise exception.PortBindingFailed(port_id=port['id']) [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 589.973686] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] [ 589.973686] env[63418]: INFO nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Terminating instance [ 590.026101] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.344s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.028724] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.354s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.030784] env[63418]: INFO nova.compute.claims [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.049865] env[63418]: INFO nova.scheduler.client.report [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Deleted allocations for instance 01deeb21-0e27-497f-9b85-c85949a3533d [ 590.232212] env[63418]: DEBUG nova.network.neutron [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.312332] env[63418]: DEBUG nova.network.neutron [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.479383] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.479870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquiring lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.561254] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27d2d55d-fed6-419f-acd6-fa9b7843c022 tempest-ServersAaction247Test-664478354 tempest-ServersAaction247Test-664478354-project-member] Lock "01deeb21-0e27-497f-9b85-c85949a3533d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.049s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.817054] env[63418]: DEBUG oslo_concurrency.lockutils [req-a0bfd929-f615-4bc4-baca-b258baefc969 req-210c226d-88dd-4777-9ae4-06fd82b474d8 service nova] Releasing lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.817054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquired lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.817285] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.349854] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.440836] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.467669] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a9b0c0-5010-4fd2-8042-831eafbeeb75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.476457] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce36327a-e26c-4723-a438-20ac45d32181 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.507212] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251bf213-27ac-4e22-add6-716fa6e762f7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.515024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb84d2df-1d32-4752-98bb-761755e77344 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.528646] env[63418]: DEBUG nova.compute.provider_tree [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.743254] env[63418]: DEBUG nova.compute.manager [req-0896a050-6bae-44bc-b1b3-a33e3b182abf req-24d83928-95aa-49dc-ae9e-12c266958c85 service nova] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Received event network-vif-deleted-e2d74ac4-1808-4ebe-ac7e-3c64149c766c {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 591.944427] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Releasing lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.944845] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 591.945045] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.945349] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0a32410-85b1-40ae-9215-bf6a7a2be0d3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.955775] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d61ea03-1d06-4d83-9aa2-e8b7cd096b08 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.982056] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a0584abb-a226-4f53-b01f-faee7ffcbf48 could not be found. [ 591.982254] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 591.982394] env[63418]: INFO nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Took 0.04 seconds to destroy the instance on the hypervisor. [ 591.982619] env[63418]: DEBUG oslo.service.loopingcall [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.982811] env[63418]: DEBUG nova.compute.manager [-] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 591.982896] env[63418]: DEBUG nova.network.neutron [-] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.012204] env[63418]: DEBUG nova.network.neutron [-] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.033943] env[63418]: DEBUG nova.scheduler.client.report [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 592.516184] env[63418]: DEBUG nova.network.neutron [-] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.539750] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.540301] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 592.544921] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.779s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.022648] env[63418]: INFO nova.compute.manager [-] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Took 1.04 seconds to deallocate network for instance. [ 593.023515] env[63418]: DEBUG nova.compute.claims [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.023858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.049026] env[63418]: DEBUG nova.compute.utils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.049026] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 593.049026] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 593.141401] env[63418]: DEBUG nova.policy [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83c6ebfffc204cb8ad6e61c5f4569017', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3c5f878ad3b4aae9da1afcdbdba4ec8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.535069] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2823bc3-7893-4693-9d57-69f9d5ba659f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.545231] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6852aea2-21a3-463d-bc8c-f9a7639acfde {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.552766] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 593.582616] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd299512-0f59-4edb-ac0e-86b994591827 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.591377] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b30889-1b1e-4760-b9e5-fc933eb3733a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.605788] env[63418]: DEBUG nova.compute.provider_tree [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.752940] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Successfully created port: e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.110873] env[63418]: DEBUG nova.scheduler.client.report [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 594.589506] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 594.615598] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.071s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.618141] env[63418]: ERROR nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Traceback (most recent call last): [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self.driver.spawn(context, instance, image_meta, [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] vm_ref = self.build_virtual_machine(instance, [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.618141] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] for vif in network_info: [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return self._sync_wrapper(fn, *args, **kwargs) [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self.wait() [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self[:] = self._gt.wait() [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return self._exit_event.wait() [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] result = hub.switch() [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.618470] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return self.greenlet.switch() [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] result = function(*args, **kwargs) [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] return func(*args, **kwargs) [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] raise e [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] nwinfo = self.network_api.allocate_for_instance( [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] created_port_ids = self._update_ports_for_instance( [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] with excutils.save_and_reraise_exception(): [ 594.621738] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] self.force_reraise() [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] raise self.value [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] updated_port = self._update_port( [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] _ensure_no_port_binding_failure(port) [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] raise exception.PortBindingFailed(port_id=port['id']) [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] nova.exception.PortBindingFailed: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. [ 594.622204] env[63418]: ERROR nova.compute.manager [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] [ 594.622490] env[63418]: DEBUG nova.compute.utils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 594.622490] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.217s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.623600] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Build of instance 3905f544-be5d-4a2a-a9c6-c7421e41b2f9 was re-scheduled: Binding failed for port 565d8ffd-fbd2-45c3-918f-51b7f9b07557, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 594.624075] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 594.624300] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquiring lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.624436] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Acquired lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.624587] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.639598] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.639741] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.639889] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.642496] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.642496] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.642496] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.643312] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.644342] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.644342] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.644342] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.644342] env[63418]: DEBUG nova.virt.hardware [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.645026] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67e6722-1b78-4bd4-a2b3-84397f280889 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.655719] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fb724d-6944-4784-87b9-f1e5fb3169b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.173822] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.345145] env[63418]: DEBUG nova.compute.manager [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Received event network-changed-e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 595.345145] env[63418]: DEBUG nova.compute.manager [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Refreshing instance network info cache due to event network-changed-e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 595.345145] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] Acquiring lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.345145] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] Acquired lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.345145] env[63418]: DEBUG nova.network.neutron [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Refreshing network info cache for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.426713] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.700370] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d484c84e-4fe4-4df3-af0f-ff4db76db84a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.708478] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66e3479-6453-4899-9644-22e947946141 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.739965] env[63418]: ERROR nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 595.739965] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.739965] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.739965] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.739965] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.739965] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.739965] env[63418]: ERROR nova.compute.manager raise self.value [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.739965] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.739965] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.739965] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.740416] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.740416] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.740416] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 595.740416] env[63418]: ERROR nova.compute.manager [ 595.740416] env[63418]: Traceback (most recent call last): [ 595.740416] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.740416] env[63418]: listener.cb(fileno) [ 595.740416] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.740416] env[63418]: result = function(*args, **kwargs) [ 595.740416] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.740416] env[63418]: return func(*args, **kwargs) [ 595.740416] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 595.740416] env[63418]: raise e [ 595.740416] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.740416] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 595.740416] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.740416] env[63418]: created_port_ids = self._update_ports_for_instance( [ 595.740416] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.740416] env[63418]: with excutils.save_and_reraise_exception(): [ 595.740416] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.740416] env[63418]: self.force_reraise() [ 595.740416] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.740416] env[63418]: raise self.value [ 595.740416] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.740416] env[63418]: updated_port = self._update_port( [ 595.740416] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.740416] env[63418]: _ensure_no_port_binding_failure(port) [ 595.740416] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.740416] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.741190] env[63418]: nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 595.741190] env[63418]: Removing descriptor: 15 [ 595.741190] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0202fbee-17f7-4c96-9d4c-a918930a8ea7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.744060] env[63418]: ERROR nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Traceback (most recent call last): [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] yield resources [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self.driver.spawn(context, instance, image_meta, [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] vm_ref = self.build_virtual_machine(instance, [ 595.744060] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] for vif in network_info: [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return self._sync_wrapper(fn, *args, **kwargs) [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self.wait() [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self[:] = self._gt.wait() [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return self._exit_event.wait() [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.744372] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] result = hub.switch() [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return self.greenlet.switch() [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] result = function(*args, **kwargs) [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return func(*args, **kwargs) [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] raise e [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] nwinfo = self.network_api.allocate_for_instance( [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] created_port_ids = self._update_ports_for_instance( [ 595.744634] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] with excutils.save_and_reraise_exception(): [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self.force_reraise() [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] raise self.value [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] updated_port = self._update_port( [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] _ensure_no_port_binding_failure(port) [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] raise exception.PortBindingFailed(port_id=port['id']) [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 595.744936] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] [ 595.745332] env[63418]: INFO nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Terminating instance [ 595.752514] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4337805-1915-441f-8206-de884102426a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.766555] env[63418]: DEBUG nova.compute.provider_tree [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.879349] env[63418]: DEBUG nova.network.neutron [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.931657] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Releasing lock "refresh_cache-3905f544-be5d-4a2a-a9c6-c7421e41b2f9" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.931909] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 595.932298] env[63418]: DEBUG nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 595.932298] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.957149] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.009602] env[63418]: DEBUG nova.network.neutron [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.250862] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.270788] env[63418]: DEBUG nova.scheduler.client.report [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 596.460650] env[63418]: DEBUG nova.network.neutron [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.512715] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e455dd-8fee-4a5c-adb6-92e2d0f61405 req-d519bd31-61f1-480b-bf5c-c78cc8223ed1 service nova] Releasing lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.513147] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquired lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.513353] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.780125] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.157s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.780125] env[63418]: ERROR nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Traceback (most recent call last): [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self.driver.spawn(context, instance, image_meta, [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.780125] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] vm_ref = self.build_virtual_machine(instance, [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] for vif in network_info: [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return self._sync_wrapper(fn, *args, **kwargs) [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self.wait() [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self[:] = self._gt.wait() [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return self._exit_event.wait() [ 596.780593] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] result = hub.switch() [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return self.greenlet.switch() [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] result = function(*args, **kwargs) [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] return func(*args, **kwargs) [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] raise e [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] nwinfo = self.network_api.allocate_for_instance( [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.780928] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] created_port_ids = self._update_ports_for_instance( [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] with excutils.save_and_reraise_exception(): [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] self.force_reraise() [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] raise self.value [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] updated_port = self._update_port( [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] _ensure_no_port_binding_failure(port) [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.781245] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] raise exception.PortBindingFailed(port_id=port['id']) [ 596.781525] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] nova.exception.PortBindingFailed: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. [ 596.781525] env[63418]: ERROR nova.compute.manager [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] [ 596.781525] env[63418]: DEBUG nova.compute.utils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.785919] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.488s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.785919] env[63418]: INFO nova.compute.claims [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.787107] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Build of instance 67b5708b-8d02-4816-9455-ea3d9414998a was re-scheduled: Binding failed for port 87d9cffe-0dab-4290-91c9-890c7f2d4f32, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 596.787756] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 596.788110] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquiring lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.788350] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Acquired lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.788593] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.888648] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "96183939-654d-4510-adf2-303f41f94640" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.888648] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "96183939-654d-4510-adf2-303f41f94640" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.964197] env[63418]: INFO nova.compute.manager [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] [instance: 3905f544-be5d-4a2a-a9c6-c7421e41b2f9] Took 1.03 seconds to deallocate network for instance. [ 597.040859] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.095173] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.313244] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.415734] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.599007] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Releasing lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.599447] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 597.599640] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.599948] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3daa7b8a-bf98-45ed-9e20-edb035988a85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.616283] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299f568a-67cd-4777-a6f6-5aed6670d241 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.646797] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37bf3e05-e797-4437-a959-96a92d4ad4e4 could not be found. [ 597.647086] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.647283] env[63418]: INFO nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 597.647524] env[63418]: DEBUG oslo.service.loopingcall [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.647732] env[63418]: DEBUG nova.compute.manager [-] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 597.647824] env[63418]: DEBUG nova.network.neutron [-] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.678140] env[63418]: DEBUG nova.network.neutron [-] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.807608] env[63418]: DEBUG nova.compute.manager [req-2e73b4b0-557c-40f9-b1b0-9f0fbfea7231 req-9f697dd7-c138-49c3-a6c1-bd3b273a9486 service nova] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Received event network-vif-deleted-e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 597.894790] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquiring lock "2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.895020] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Lock "2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.918225] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Releasing lock "refresh_cache-67b5708b-8d02-4816-9455-ea3d9414998a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.918466] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 597.918650] env[63418]: DEBUG nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 597.918814] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.943903] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.997508] env[63418]: INFO nova.scheduler.client.report [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Deleted allocations for instance 3905f544-be5d-4a2a-a9c6-c7421e41b2f9 [ 598.124426] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "e7a4d56b-f785-44b4-acb3-6390a4bad83c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.124697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "e7a4d56b-f785-44b4-acb3-6390a4bad83c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.180962] env[63418]: DEBUG nova.network.neutron [-] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.276730] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266668b6-13c2-4925-9130-3680444c4fbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.284357] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca3b45c-0c76-4aef-af50-fa17aac1b2f1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.320883] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb956e60-8f04-491e-83e0-04aab9208324 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.328142] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4546c2-96bc-499d-a6aa-adbb98415f7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.341439] env[63418]: DEBUG nova.compute.provider_tree [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.448293] env[63418]: DEBUG nova.network.neutron [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.510123] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee77424d-dc05-478f-a71f-34fab2df76ab tempest-ServersTestJSON-49657505 tempest-ServersTestJSON-49657505-project-member] Lock "3905f544-be5d-4a2a-a9c6-c7421e41b2f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.470s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.690235] env[63418]: INFO nova.compute.manager [-] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Took 1.04 seconds to deallocate network for instance. [ 598.691644] env[63418]: DEBUG nova.compute.claims [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.692047] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.848022] env[63418]: DEBUG nova.scheduler.client.report [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 598.950947] env[63418]: INFO nova.compute.manager [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] [instance: 67b5708b-8d02-4816-9455-ea3d9414998a] Took 1.03 seconds to deallocate network for instance. [ 599.012716] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 599.353846] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.354465] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 599.357093] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.332s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.357281] env[63418]: DEBUG nova.objects.instance [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 599.540031] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.865338] env[63418]: DEBUG nova.compute.utils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.869497] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 599.869675] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.979019] env[63418]: DEBUG nova.policy [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32828aef745b4233b82b35a86a17eb50', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f60d0670fb044c5812915598a8703bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.993968] env[63418]: INFO nova.scheduler.client.report [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Deleted allocations for instance 67b5708b-8d02-4816-9455-ea3d9414998a [ 600.370339] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 600.374849] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e6935587-1fcc-481a-84fd-b490f8ec63c8 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.377189] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.753s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.482271] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Successfully created port: d6b6de70-fcba-4f83-8d0b-58583df82063 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.505995] env[63418]: DEBUG oslo_concurrency.lockutils [None req-77ccd90d-0c0b-4734-8b18-9b9f8f33020e tempest-ServersWithSpecificFlavorTestJSON-1533174611 tempest-ServersWithSpecificFlavorTestJSON-1533174611-project-member] Lock "67b5708b-8d02-4816-9455-ea3d9414998a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.135s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.008886] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 601.285289] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c96b40-cf16-439c-ac32-45d390f7b3b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.293848] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0eae5fc-8461-471d-8c2b-50de5103c7de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.331098] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd476ea8-630b-4873-b210-a78300994aff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.340994] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7fcc155-9076-49ad-95a8-449a52c47a5b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.357421] env[63418]: DEBUG nova.compute.provider_tree [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.362133] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 601.362609] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 601.390089] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 601.422156] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.422649] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.422828] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.423068] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.423847] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.424078] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.424388] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.424590] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.424789] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.424982] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.425212] env[63418]: DEBUG nova.virt.hardware [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.426104] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735e81f1-c9ae-49dd-a90f-8e324409af9e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.435420] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac06c6d-27b5-42d5-8f7d-913c1d949586 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.533863] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.585897] env[63418]: DEBUG nova.compute.manager [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Received event network-changed-d6b6de70-fcba-4f83-8d0b-58583df82063 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 601.585897] env[63418]: DEBUG nova.compute.manager [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Refreshing instance network info cache due to event network-changed-d6b6de70-fcba-4f83-8d0b-58583df82063. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 601.585897] env[63418]: DEBUG oslo_concurrency.lockutils [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] Acquiring lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.585995] env[63418]: DEBUG oslo_concurrency.lockutils [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] Acquired lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.586429] env[63418]: DEBUG nova.network.neutron [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Refreshing network info cache for port d6b6de70-fcba-4f83-8d0b-58583df82063 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 601.718738] env[63418]: ERROR nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 601.718738] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.718738] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.718738] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.718738] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.718738] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.718738] env[63418]: ERROR nova.compute.manager raise self.value [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.718738] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.718738] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.718738] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.719123] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.719123] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.719123] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 601.719123] env[63418]: ERROR nova.compute.manager [ 601.719123] env[63418]: Traceback (most recent call last): [ 601.719123] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.719123] env[63418]: listener.cb(fileno) [ 601.719123] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.719123] env[63418]: result = function(*args, **kwargs) [ 601.719123] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.719123] env[63418]: return func(*args, **kwargs) [ 601.719123] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.719123] env[63418]: raise e [ 601.719123] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.719123] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 601.719123] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.719123] env[63418]: created_port_ids = self._update_ports_for_instance( [ 601.719123] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.719123] env[63418]: with excutils.save_and_reraise_exception(): [ 601.719123] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.719123] env[63418]: self.force_reraise() [ 601.719123] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.719123] env[63418]: raise self.value [ 601.719123] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.719123] env[63418]: updated_port = self._update_port( [ 601.719123] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.719123] env[63418]: _ensure_no_port_binding_failure(port) [ 601.719123] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.719123] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.719739] env[63418]: nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 601.719739] env[63418]: Removing descriptor: 15 [ 601.719739] env[63418]: ERROR nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Traceback (most recent call last): [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] yield resources [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self.driver.spawn(context, instance, image_meta, [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.719739] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] vm_ref = self.build_virtual_machine(instance, [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] for vif in network_info: [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return self._sync_wrapper(fn, *args, **kwargs) [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self.wait() [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self[:] = self._gt.wait() [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return self._exit_event.wait() [ 601.719998] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] result = hub.switch() [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return self.greenlet.switch() [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] result = function(*args, **kwargs) [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return func(*args, **kwargs) [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] raise e [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] nwinfo = self.network_api.allocate_for_instance( [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.720380] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] created_port_ids = self._update_ports_for_instance( [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] with excutils.save_and_reraise_exception(): [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self.force_reraise() [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] raise self.value [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] updated_port = self._update_port( [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] _ensure_no_port_binding_failure(port) [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.720659] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] raise exception.PortBindingFailed(port_id=port['id']) [ 601.720915] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 601.720915] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] [ 601.720915] env[63418]: INFO nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Terminating instance [ 601.864953] env[63418]: DEBUG nova.scheduler.client.report [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 601.873851] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 601.873851] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 601.874777] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Rebuilding the list of instances to heal {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10282}} [ 602.112331] env[63418]: DEBUG nova.network.neutron [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.175578] env[63418]: DEBUG nova.network.neutron [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.225467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquiring lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.371916] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.373351] env[63418]: ERROR nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] Traceback (most recent call last): [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self.driver.spawn(context, instance, image_meta, [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] vm_ref = self.build_virtual_machine(instance, [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.373351] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] for vif in network_info: [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] return self._sync_wrapper(fn, *args, **kwargs) [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self.wait() [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self[:] = self._gt.wait() [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] return self._exit_event.wait() [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] current.throw(*self._exc) [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.377061] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] result = function(*args, **kwargs) [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] return func(*args, **kwargs) [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] raise e [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] nwinfo = self.network_api.allocate_for_instance( [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] created_port_ids = self._update_ports_for_instance( [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] with excutils.save_and_reraise_exception(): [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] self.force_reraise() [ 602.377510] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] raise self.value [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] updated_port = self._update_port( [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] _ensure_no_port_binding_failure(port) [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] raise exception.PortBindingFailed(port_id=port['id']) [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] nova.exception.PortBindingFailed: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. [ 602.377905] env[63418]: ERROR nova.compute.manager [instance: 58850986-3ebf-465f-b832-9b245742a34a] [ 602.377905] env[63418]: DEBUG nova.compute.utils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.378199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.142s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.378199] env[63418]: DEBUG nova.objects.instance [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lazy-loading 'resources' on Instance uuid ed637299-7e69-43d7-85f9-8fcbd6e90dec {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 602.378199] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Build of instance 58850986-3ebf-465f-b832-9b245742a34a was re-scheduled: Binding failed for port c5dfed01-0c53-473a-9dad-48f1650256e8, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 602.378199] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 602.378336] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquiring lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.378336] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Acquired lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.378336] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.382276] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 602.382588] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 602.382867] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 602.382925] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 602.384062] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 602.384062] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 602.417173] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "refresh_cache-7e43d259-f361-43d8-8f03-72b303680478" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.417327] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquired lock "refresh_cache-7e43d259-f361-43d8-8f03-72b303680478" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.417472] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Forcefully refreshing network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 602.417621] env[63418]: DEBUG nova.objects.instance [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lazy-loading 'info_cache' on Instance uuid 7e43d259-f361-43d8-8f03-72b303680478 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 602.677767] env[63418]: DEBUG oslo_concurrency.lockutils [req-33c57f8a-7b96-436e-97e7-2c2d6a0ca972 req-1508b102-7df0-47fa-b154-633f16e76d0c service nova] Releasing lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.678204] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquired lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.678378] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.904725] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.978640] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.201020] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.287713] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.336521] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676573bb-e587-4e15-89c0-68796136bb04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.349223] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67d403b-e7ed-4792-9ecb-6f19f5690f6f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.384296] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab457e5-ee31-4cb2-8fea-f7ab253d1827 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.391796] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3915b4-e364-47c0-8cfc-2d7f72e89006 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.405158] env[63418]: DEBUG nova.compute.provider_tree [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.451218] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.481396] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Releasing lock "refresh_cache-58850986-3ebf-465f-b832-9b245742a34a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.481620] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 603.481781] env[63418]: DEBUG nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.482070] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.512749] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.760888] env[63418]: DEBUG nova.compute.manager [req-5b8293d5-7e06-4f5b-b2dc-419bbd0e0aaf req-0f9d965b-ac37-4a69-a71f-9e7498b9aa76 service nova] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Received event network-vif-deleted-d6b6de70-fcba-4f83-8d0b-58583df82063 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 603.790219] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Releasing lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.790676] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 603.790866] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.791248] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-297a8182-b384-4a16-90a6-80f1c037e756 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.800774] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d948a57-6a69-405f-b25a-12830453ea22 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.822293] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9b2d0dfa-6820-451f-b86d-590f89928278 could not be found. [ 603.822552] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.822742] env[63418]: INFO nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Took 0.03 seconds to destroy the instance on the hypervisor. [ 603.822986] env[63418]: DEBUG oslo.service.loopingcall [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.823227] env[63418]: DEBUG nova.compute.manager [-] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.823343] env[63418]: DEBUG nova.network.neutron [-] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.849836] env[63418]: DEBUG nova.network.neutron [-] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.908674] env[63418]: DEBUG nova.scheduler.client.report [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 604.017577] env[63418]: DEBUG nova.network.neutron [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.032572] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.352870] env[63418]: DEBUG nova.network.neutron [-] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.414932] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.417602] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.443s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.417950] env[63418]: DEBUG nova.objects.instance [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lazy-loading 'resources' on Instance uuid a8c7f192-4672-43cc-8c38-0c33ce633765 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 604.443603] env[63418]: INFO nova.scheduler.client.report [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Deleted allocations for instance ed637299-7e69-43d7-85f9-8fcbd6e90dec [ 604.522218] env[63418]: INFO nova.compute.manager [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] [instance: 58850986-3ebf-465f-b832-9b245742a34a] Took 1.04 seconds to deallocate network for instance. [ 604.535243] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Releasing lock "refresh_cache-7e43d259-f361-43d8-8f03-72b303680478" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.535480] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Updated the network info_cache for instance {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10349}} [ 604.535702] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.536858] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.537263] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.537263] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.537385] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.537529] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.537661] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 604.537796] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.859320] env[63418]: INFO nova.compute.manager [-] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Took 1.04 seconds to deallocate network for instance. [ 604.861845] env[63418]: DEBUG nova.compute.claims [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.862090] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.960188] env[63418]: DEBUG oslo_concurrency.lockutils [None req-11bccfee-3d1c-42aa-aced-96c7c4610b9b tempest-ServerDiagnosticsV248Test-1099631268 tempest-ServerDiagnosticsV248Test-1099631268-project-member] Lock "ed637299-7e69-43d7-85f9-8fcbd6e90dec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.008s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.041090] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.419964] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8cb355-af42-422c-9cf4-ecde94c9e7b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.427844] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cddbaaa-ea76-4d8d-a13b-39253fe8409a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.465615] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4292db-ab6f-4e2d-9f0c-0655624a51f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.473440] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af3d319-7aab-440f-8714-579b048ad9b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.487227] env[63418]: DEBUG nova.compute.provider_tree [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.562204] env[63418]: INFO nova.scheduler.client.report [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Deleted allocations for instance 58850986-3ebf-465f-b832-9b245742a34a [ 605.992224] env[63418]: DEBUG nova.scheduler.client.report [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 606.072076] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b0d5f0ef-aa99-4e9f-b851-1ea27fafe42d tempest-ServerDiagnosticsTest-1322591085 tempest-ServerDiagnosticsTest-1322591085-project-member] Lock "58850986-3ebf-465f-b832-9b245742a34a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.825s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.506934] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.089s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.510185] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.966s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.511436] env[63418]: DEBUG nova.objects.instance [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lazy-loading 'resources' on Instance uuid 7e43d259-f361-43d8-8f03-72b303680478 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 606.536405] env[63418]: INFO nova.scheduler.client.report [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Deleted allocations for instance a8c7f192-4672-43cc-8c38-0c33ce633765 [ 606.575887] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 606.773527] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquiring lock "702be472-a4bc-4089-94de-4355ee00684a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.773753] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Lock "702be472-a4bc-4089-94de-4355ee00684a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.048916] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fea0d713-b74c-4157-80b3-65f14867f491 tempest-ServerShowV257Test-439314089 tempest-ServerShowV257Test-439314089-project-member] Lock "a8c7f192-4672-43cc-8c38-0c33ce633765" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.304s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.118754] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.553891] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca3900e-af20-43eb-ac4c-6a2a212c187f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.563436] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e4619a-24b7-45e1-bed7-31cc6f5f39f3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.602082] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa4e8c4-03a0-49ee-8d38-145b96257726 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.609839] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4486f9e2-dabc-4500-bb23-9431276ebc67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.624585] env[63418]: DEBUG nova.compute.provider_tree [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.127946] env[63418]: DEBUG nova.scheduler.client.report [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 608.634970] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.637166] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.692s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.678130] env[63418]: INFO nova.scheduler.client.report [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Deleted allocations for instance 7e43d259-f361-43d8-8f03-72b303680478 [ 609.190975] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ae977f86-cade-42ff-a679-f320d3fad315 tempest-ServerShowV247Test-1948825370 tempest-ServerShowV247Test-1948825370-project-member] Lock "7e43d259-f361-43d8-8f03-72b303680478" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.977s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.622590] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2e5175-96bf-4163-aa17-7d802f879f09 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.630491] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd027b7-8f70-44f4-9f9b-41ddd9d457a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.666090] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae64058b-3d10-44d2-a8e2-73ccb8365dd0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.672035] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c59a06c-adea-424a-84fa-04f96191ab9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.686742] env[63418]: DEBUG nova.compute.provider_tree [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.190198] env[63418]: DEBUG nova.scheduler.client.report [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 610.700338] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.700968] env[63418]: ERROR nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Traceback (most recent call last): [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self.driver.spawn(context, instance, image_meta, [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] vm_ref = self.build_virtual_machine(instance, [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.700968] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] for vif in network_info: [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] return self._sync_wrapper(fn, *args, **kwargs) [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self.wait() [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self[:] = self._gt.wait() [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] return self._exit_event.wait() [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] current.throw(*self._exc) [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.702321] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] result = function(*args, **kwargs) [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] return func(*args, **kwargs) [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] raise e [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] nwinfo = self.network_api.allocate_for_instance( [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] created_port_ids = self._update_ports_for_instance( [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] with excutils.save_and_reraise_exception(): [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] self.force_reraise() [ 610.703678] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] raise self.value [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] updated_port = self._update_port( [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] _ensure_no_port_binding_failure(port) [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] raise exception.PortBindingFailed(port_id=port['id']) [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] nova.exception.PortBindingFailed: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. [ 610.704345] env[63418]: ERROR nova.compute.manager [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] [ 610.704345] env[63418]: DEBUG nova.compute.utils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.708176] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.324s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.710351] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Build of instance 34cf7345-c244-44bd-aa1f-4788ae8f6da5 was re-scheduled: Binding failed for port d735e501-e398-41b4-a4f6-39af715bd5c3, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 610.710826] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 610.711064] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.711206] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.711355] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.911924] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "6b81dbe2-aa9e-4561-962b-2af167234b90" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.911924] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.241647] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.353327] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "17a055e4-13da-4984-ae64-af1ae9d652f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.353506] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.486621] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.727498] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9002a913-19a7-4f0e-9101-e32593bb32af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.735736] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5886687e-6a87-44aa-a418-e1d941d0e866 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.767636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08494b6a-ca22-47ab-9a79-7ca5be1ce009 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.776050] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70a7b2d-8975-41ef-8e31-74a38d3ba227 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.789886] env[63418]: DEBUG nova.compute.provider_tree [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.968307] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.968562] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.989575] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-34cf7345-c244-44bd-aa1f-4788ae8f6da5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.989947] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 611.993052] env[63418]: DEBUG nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 611.993052] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.020531] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.293301] env[63418]: DEBUG nova.scheduler.client.report [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 612.524426] env[63418]: DEBUG nova.network.neutron [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.800693] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.800693] env[63418]: ERROR nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Traceback (most recent call last): [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self.driver.spawn(context, instance, image_meta, [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.800693] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] vm_ref = self.build_virtual_machine(instance, [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] for vif in network_info: [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] return self._sync_wrapper(fn, *args, **kwargs) [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self.wait() [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self[:] = self._gt.wait() [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] return self._exit_event.wait() [ 612.801597] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] current.throw(*self._exc) [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] result = function(*args, **kwargs) [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] return func(*args, **kwargs) [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] raise e [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] nwinfo = self.network_api.allocate_for_instance( [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] created_port_ids = self._update_ports_for_instance( [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.802241] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] with excutils.save_and_reraise_exception(): [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] self.force_reraise() [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] raise self.value [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] updated_port = self._update_port( [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] _ensure_no_port_binding_failure(port) [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] raise exception.PortBindingFailed(port_id=port['id']) [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] nova.exception.PortBindingFailed: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. [ 612.802565] env[63418]: ERROR nova.compute.manager [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] [ 612.802884] env[63418]: DEBUG nova.compute.utils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.806746] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Build of instance 5caab97d-833e-4356-af51-4a991d0e5b44 was re-scheduled: Binding failed for port ae13bf0a-29af-4ffc-a1e0-23ecd8d259b4, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 612.806746] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 612.806746] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquiring lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.806746] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Acquired lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.807085] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.807085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.871s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.028748] env[63418]: INFO nova.compute.manager [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 34cf7345-c244-44bd-aa1f-4788ae8f6da5] Took 1.04 seconds to deallocate network for instance. [ 613.365502] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.543090] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.866144] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1e545b-e695-4988-a7ad-45f9ea5edd02 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.878192] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dbe3b2-48d0-45b5-a753-996e694bdb98 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.929727] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad46df5-d291-4e60-b129-bf3f9a94e2b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.951896] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd254f38-eff7-4e60-a404-aa6538d48a18 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.973274] env[63418]: DEBUG nova.compute.provider_tree [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.051035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Releasing lock "refresh_cache-5caab97d-833e-4356-af51-4a991d0e5b44" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.051839] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 614.052184] env[63418]: DEBUG nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 614.052606] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.072937] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.077503] env[63418]: INFO nova.scheduler.client.report [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocations for instance 34cf7345-c244-44bd-aa1f-4788ae8f6da5 [ 614.477534] env[63418]: DEBUG nova.scheduler.client.report [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 614.578070] env[63418]: DEBUG nova.network.neutron [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.592363] env[63418]: DEBUG oslo_concurrency.lockutils [None req-75596539-b64a-4a3e-a07a-5446fce813c9 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "34cf7345-c244-44bd-aa1f-4788ae8f6da5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.435s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.986645] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.178s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.986645] env[63418]: ERROR nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Traceback (most recent call last): [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self.driver.spawn(context, instance, image_meta, [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.986645] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] vm_ref = self.build_virtual_machine(instance, [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] for vif in network_info: [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] return self._sync_wrapper(fn, *args, **kwargs) [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self.wait() [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self[:] = self._gt.wait() [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] return self._exit_event.wait() [ 614.987290] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] current.throw(*self._exc) [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] result = function(*args, **kwargs) [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] return func(*args, **kwargs) [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] raise e [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] nwinfo = self.network_api.allocate_for_instance( [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] created_port_ids = self._update_ports_for_instance( [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.987694] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] with excutils.save_and_reraise_exception(): [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] self.force_reraise() [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] raise self.value [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] updated_port = self._update_port( [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] _ensure_no_port_binding_failure(port) [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] raise exception.PortBindingFailed(port_id=port['id']) [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] nova.exception.PortBindingFailed: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. [ 614.988057] env[63418]: ERROR nova.compute.manager [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] [ 614.988391] env[63418]: DEBUG nova.compute.utils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.988627] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.508s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.995096] env[63418]: INFO nova.compute.claims [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.998534] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Build of instance b1b24de1-3d7f-40b7-9183-c4b1f35f8c94 was re-scheduled: Binding failed for port f9ebb350-3730-4f9d-b221-b6af2642a7b2, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 614.999029] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 615.000195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquiring lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.000394] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Acquired lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.000559] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.081569] env[63418]: INFO nova.compute.manager [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] [instance: 5caab97d-833e-4356-af51-4a991d0e5b44] Took 1.03 seconds to deallocate network for instance. [ 615.096431] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 615.523661] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.619746] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.622896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.669080] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.669347] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.115342] env[63418]: INFO nova.scheduler.client.report [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Deleted allocations for instance 5caab97d-833e-4356-af51-4a991d0e5b44 [ 616.121854] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Releasing lock "refresh_cache-b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.122094] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 616.122275] env[63418]: DEBUG nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 616.122432] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.152658] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.499129] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa0ed45-79b8-44fb-8be4-d25b5e4bbcea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.507407] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd441b66-5523-4225-8f9c-127687ad2b73 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.542021] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f939b7df-f8d1-409b-825d-ab6c6b65393d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.551164] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef75a1c1-ce41-4482-812b-10ab57441281 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.566948] env[63418]: DEBUG nova.compute.provider_tree [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.629573] env[63418]: DEBUG oslo_concurrency.lockutils [None req-19b7f214-540f-4c21-8fc0-80a44c85a6da tempest-ImagesOneServerTestJSON-1964391107 tempest-ImagesOneServerTestJSON-1964391107-project-member] Lock "5caab97d-833e-4356-af51-4a991d0e5b44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.070s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.654356] env[63418]: DEBUG nova.network.neutron [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.676761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.676982] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.070480] env[63418]: DEBUG nova.scheduler.client.report [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 617.137210] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 617.157493] env[63418]: INFO nova.compute.manager [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] [instance: b1b24de1-3d7f-40b7-9183-c4b1f35f8c94] Took 1.03 seconds to deallocate network for instance. [ 617.576475] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.577173] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 617.579979] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.556s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.665912] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.858220] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.858496] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.088485] env[63418]: DEBUG nova.compute.utils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.093243] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 618.093243] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 618.160857] env[63418]: DEBUG nova.policy [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fa569d01e34f4d68a108a276345343b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca98e5f845c64f07abb2559cd10efe1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.185911] env[63418]: INFO nova.scheduler.client.report [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Deleted allocations for instance b1b24de1-3d7f-40b7-9183-c4b1f35f8c94 [ 618.544355] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Successfully created port: 6272afaa-e1e6-4386-82ea-26eb9aba171e {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.594536] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 618.633201] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39dd0301-e608-4834-b376-b481864ae15a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.643501] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fe5fa0-b161-400c-9056-aea2bc69450f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.681645] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa2d6719-019e-4e25-9dfe-5a24672587e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.690859] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddcdb37-e24f-4cd2-ab8f-0b61a15ff71e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.707078] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8a3cd89c-12a9-40ca-b9c9-83b1963a9db3 tempest-ServerMetadataTestJSON-2027739093 tempest-ServerMetadataTestJSON-2027739093-project-member] Lock "b1b24de1-3d7f-40b7-9183-c4b1f35f8c94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.311s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.707764] env[63418]: DEBUG nova.compute.provider_tree [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.212967] env[63418]: DEBUG nova.scheduler.client.report [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 619.216402] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 619.245595] env[63418]: DEBUG nova.compute.manager [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Received event network-changed-6272afaa-e1e6-4386-82ea-26eb9aba171e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 619.245805] env[63418]: DEBUG nova.compute.manager [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Refreshing instance network info cache due to event network-changed-6272afaa-e1e6-4386-82ea-26eb9aba171e. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 619.246059] env[63418]: DEBUG oslo_concurrency.lockutils [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] Acquiring lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.246162] env[63418]: DEBUG oslo_concurrency.lockutils [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] Acquired lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.246280] env[63418]: DEBUG nova.network.neutron [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Refreshing network info cache for port 6272afaa-e1e6-4386-82ea-26eb9aba171e {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.470829] env[63418]: ERROR nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 619.470829] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.470829] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.470829] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.470829] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.470829] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.470829] env[63418]: ERROR nova.compute.manager raise self.value [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.470829] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.470829] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.470829] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.471352] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.471352] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.471352] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 619.471352] env[63418]: ERROR nova.compute.manager [ 619.471352] env[63418]: Traceback (most recent call last): [ 619.471352] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.471352] env[63418]: listener.cb(fileno) [ 619.471352] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.471352] env[63418]: result = function(*args, **kwargs) [ 619.471352] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.471352] env[63418]: return func(*args, **kwargs) [ 619.471352] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 619.471352] env[63418]: raise e [ 619.471352] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.471352] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 619.471352] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.471352] env[63418]: created_port_ids = self._update_ports_for_instance( [ 619.471352] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.471352] env[63418]: with excutils.save_and_reraise_exception(): [ 619.471352] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.471352] env[63418]: self.force_reraise() [ 619.471352] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.471352] env[63418]: raise self.value [ 619.471352] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.471352] env[63418]: updated_port = self._update_port( [ 619.471352] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.471352] env[63418]: _ensure_no_port_binding_failure(port) [ 619.471352] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.471352] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.472163] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 619.472163] env[63418]: Removing descriptor: 15 [ 619.606723] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 619.631578] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.631818] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.631971] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.632563] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.632563] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.632563] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.632699] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.632782] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.632944] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.633117] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.633318] env[63418]: DEBUG nova.virt.hardware [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.634245] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c045b8e1-6efc-426b-8724-a62f3c3183fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.642796] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b214ddcd-becc-40f6-b9ef-eac785d8fc4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.655713] env[63418]: ERROR nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Traceback (most recent call last): [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] yield resources [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self.driver.spawn(context, instance, image_meta, [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] vm_ref = self.build_virtual_machine(instance, [ 619.655713] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] for vif in network_info: [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] return self._sync_wrapper(fn, *args, **kwargs) [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self.wait() [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self[:] = self._gt.wait() [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] return self._exit_event.wait() [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 619.656042] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] current.throw(*self._exc) [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] result = function(*args, **kwargs) [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] return func(*args, **kwargs) [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] raise e [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] nwinfo = self.network_api.allocate_for_instance( [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] created_port_ids = self._update_ports_for_instance( [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] with excutils.save_and_reraise_exception(): [ 619.656358] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self.force_reraise() [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] raise self.value [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] updated_port = self._update_port( [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] _ensure_no_port_binding_failure(port) [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] raise exception.PortBindingFailed(port_id=port['id']) [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 619.656657] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] [ 619.656657] env[63418]: INFO nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Terminating instance [ 619.724727] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.145s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.725345] env[63418]: ERROR nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Traceback (most recent call last): [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self.driver.spawn(context, instance, image_meta, [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] vm_ref = self.build_virtual_machine(instance, [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.725345] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] for vif in network_info: [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] return self._sync_wrapper(fn, *args, **kwargs) [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self.wait() [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self[:] = self._gt.wait() [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] return self._exit_event.wait() [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] current.throw(*self._exc) [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.725647] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] result = function(*args, **kwargs) [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] return func(*args, **kwargs) [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] raise e [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] nwinfo = self.network_api.allocate_for_instance( [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] created_port_ids = self._update_ports_for_instance( [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] with excutils.save_and_reraise_exception(): [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] self.force_reraise() [ 619.725949] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] raise self.value [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] updated_port = self._update_port( [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] _ensure_no_port_binding_failure(port) [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] raise exception.PortBindingFailed(port_id=port['id']) [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] nova.exception.PortBindingFailed: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. [ 619.726262] env[63418]: ERROR nova.compute.manager [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] [ 619.726262] env[63418]: DEBUG nova.compute.utils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.727234] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.035s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.732504] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Build of instance a0584abb-a226-4f53-b01f-faee7ffcbf48 was re-scheduled: Binding failed for port e2d74ac4-1808-4ebe-ac7e-3c64149c766c, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 619.734070] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 619.734070] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquiring lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.734070] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Acquired lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.734070] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.770400] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.774679] env[63418]: DEBUG nova.network.neutron [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.976148] env[63418]: DEBUG nova.network.neutron [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.160905] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquiring lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.261379] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.342235] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.478906] env[63418]: DEBUG oslo_concurrency.lockutils [req-2177d56f-f228-471a-a442-504a39bea4c9 req-4ba34e2c-092b-4e3b-94f3-432b164d4c92 service nova] Releasing lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.479590] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquired lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.479590] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.737170] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63ae10e-1213-4570-8d2c-7c0d13db081a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.745704] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27a1669-8f8e-40b7-b1ef-063b808668c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.786086] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13eeae7-e6a9-470f-8e58-ca6386da1a32 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.799360] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fdee55-8d52-4f8c-91c1-f912ab4a8892 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.814724] env[63418]: DEBUG nova.compute.provider_tree [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.847000] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Releasing lock "refresh_cache-a0584abb-a226-4f53-b01f-faee7ffcbf48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.847000] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 620.847000] env[63418]: DEBUG nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 620.847000] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.876167] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.003994] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.100638] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.318056] env[63418]: DEBUG nova.scheduler.client.report [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 621.363076] env[63418]: DEBUG nova.compute.manager [req-2c9b5c25-760d-4338-8eb4-679fd4b9da3d req-4acba5e7-fc0b-42a3-aee0-b74bfdb5fdbf service nova] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Received event network-vif-deleted-6272afaa-e1e6-4386-82ea-26eb9aba171e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 621.381080] env[63418]: DEBUG nova.network.neutron [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.603632] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Releasing lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.605426] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 621.605426] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 621.609375] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da1d860f-2a84-4dd2-851a-b5f948e70cd9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.622715] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a487a9d2-1eda-4b49-acdd-6786b0301333 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.646592] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0c55e7b9-cb53-4525-8524-4ef62a51a89f could not be found. [ 621.647114] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.647440] env[63418]: INFO nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 621.647838] env[63418]: DEBUG oslo.service.loopingcall [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.648292] env[63418]: DEBUG nova.compute.manager [-] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 621.648531] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.819908] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.826600] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.099s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.827251] env[63418]: ERROR nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Traceback (most recent call last): [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self.driver.spawn(context, instance, image_meta, [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] vm_ref = self.build_virtual_machine(instance, [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.827251] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] for vif in network_info: [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return self._sync_wrapper(fn, *args, **kwargs) [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self.wait() [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self[:] = self._gt.wait() [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return self._exit_event.wait() [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] result = hub.switch() [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.827653] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return self.greenlet.switch() [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] result = function(*args, **kwargs) [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] return func(*args, **kwargs) [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] raise e [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] nwinfo = self.network_api.allocate_for_instance( [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] created_port_ids = self._update_ports_for_instance( [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] with excutils.save_and_reraise_exception(): [ 621.828083] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] self.force_reraise() [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] raise self.value [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] updated_port = self._update_port( [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] _ensure_no_port_binding_failure(port) [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] raise exception.PortBindingFailed(port_id=port['id']) [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] nova.exception.PortBindingFailed: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. [ 621.828486] env[63418]: ERROR nova.compute.manager [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] [ 621.828932] env[63418]: DEBUG nova.compute.utils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.829203] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.289s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.830457] env[63418]: INFO nova.compute.claims [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.835936] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Build of instance 37bf3e05-e797-4437-a959-96a92d4ad4e4 was re-scheduled: Binding failed for port e6184f2f-8e54-43f1-bfcc-27cf3d2adfa3, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 621.835936] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 621.835936] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquiring lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.836128] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Acquired lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.836163] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.883698] env[63418]: INFO nova.compute.manager [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] [instance: a0584abb-a226-4f53-b01f-faee7ffcbf48] Took 1.04 seconds to deallocate network for instance. [ 622.322445] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.359112] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.470195] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.825012] env[63418]: INFO nova.compute.manager [-] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Took 1.18 seconds to deallocate network for instance. [ 622.830302] env[63418]: DEBUG nova.compute.claims [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.830535] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.913876] env[63418]: INFO nova.scheduler.client.report [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Deleted allocations for instance a0584abb-a226-4f53-b01f-faee7ffcbf48 [ 622.977149] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Releasing lock "refresh_cache-37bf3e05-e797-4437-a959-96a92d4ad4e4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.977149] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 622.977149] env[63418]: DEBUG nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 622.977149] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.004822] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.345418] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7df09e-74f2-40ec-9c4e-d75e2eee6122 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.355744] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e62c628-6005-44d3-a940-2c67b6905a8e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.391068] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3aa118-5b7a-4414-8ee9-4437ec205208 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.398831] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4577ade1-a211-47e5-ae8e-51b3b60e9808 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.414310] env[63418]: DEBUG nova.compute.provider_tree [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.423486] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c3b0b5c-b848-4e6c-84a4-acf6913afd5b tempest-TenantUsagesTestJSON-600174989 tempest-TenantUsagesTestJSON-600174989-project-member] Lock "a0584abb-a226-4f53-b01f-faee7ffcbf48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.580s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.508804] env[63418]: DEBUG nova.network.neutron [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.920121] env[63418]: DEBUG nova.scheduler.client.report [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 623.930248] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 624.014021] env[63418]: INFO nova.compute.manager [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] [instance: 37bf3e05-e797-4437-a959-96a92d4ad4e4] Took 1.04 seconds to deallocate network for instance. [ 624.198388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "7615244b-fa19-4dbb-a851-604bb6fc2725" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.199060] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.238314] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "0c151b0c-c383-4421-b506-b7afa95e2072" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.238577] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "0c151b0c-c383-4421-b506-b7afa95e2072" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.266174] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "f632d71d-498f-4914-9895-8f37187a295f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.266477] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "f632d71d-498f-4914-9895-8f37187a295f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.429661] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.430310] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 624.434233] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.900s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.436198] env[63418]: INFO nova.compute.claims [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.464213] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.942047] env[63418]: DEBUG nova.compute.utils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.943477] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 624.943715] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.013552] env[63418]: DEBUG nova.policy [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa9ae41d400c495a8b6142d6687b8197', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f3d44332bde412693353e5a0b0ea44d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.050560] env[63418]: INFO nova.scheduler.client.report [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Deleted allocations for instance 37bf3e05-e797-4437-a959-96a92d4ad4e4 [ 625.450902] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 625.458411] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Successfully created port: 04d277c1-ffc4-4d28-bff1-4287ba200126 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.563306] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbc8a86a-ce9f-40c8-91f7-d916c944c1b8 tempest-DeleteServersAdminTestJSON-1488719117 tempest-DeleteServersAdminTestJSON-1488719117-project-member] Lock "37bf3e05-e797-4437-a959-96a92d4ad4e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.968s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.823260] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.823945] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.979965] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860e2598-c1d9-4c37-9b6e-f3629eb563ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.987869] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f8b365-0c9b-430c-9b00-a0454203c89c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.022777] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc95cac8-5bf4-437b-a902-9123b04a57a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.030820] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84f4fa9-5764-473a-a834-1f07e189ffa0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.049505] env[63418]: DEBUG nova.compute.provider_tree [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.068642] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 626.460342] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 626.494331] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.494598] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.494752] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.494929] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.495095] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.495247] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.495454] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.495639] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.495810] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.495970] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.496164] env[63418]: DEBUG nova.virt.hardware [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.497504] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8023a7d0-750e-4876-90da-7fb4a4a83ebb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.507094] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365aee10-d5a5-429f-b3e0-e8aaf49394c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.553170] env[63418]: DEBUG nova.scheduler.client.report [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 626.607340] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.966338] env[63418]: DEBUG nova.compute.manager [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Received event network-changed-04d277c1-ffc4-4d28-bff1-4287ba200126 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 626.966511] env[63418]: DEBUG nova.compute.manager [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Refreshing instance network info cache due to event network-changed-04d277c1-ffc4-4d28-bff1-4287ba200126. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 626.966843] env[63418]: DEBUG oslo_concurrency.lockutils [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] Acquiring lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.966905] env[63418]: DEBUG oslo_concurrency.lockutils [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] Acquired lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.967020] env[63418]: DEBUG nova.network.neutron [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Refreshing network info cache for port 04d277c1-ffc4-4d28-bff1-4287ba200126 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 627.062609] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.629s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.064764] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 627.071353] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.208s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.260636] env[63418]: ERROR nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 627.260636] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.260636] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.260636] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.260636] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.260636] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.260636] env[63418]: ERROR nova.compute.manager raise self.value [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.260636] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.260636] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.260636] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.261177] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.261177] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.261177] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 627.261177] env[63418]: ERROR nova.compute.manager [ 627.261177] env[63418]: Traceback (most recent call last): [ 627.261177] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.261177] env[63418]: listener.cb(fileno) [ 627.261177] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.261177] env[63418]: result = function(*args, **kwargs) [ 627.261177] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.261177] env[63418]: return func(*args, **kwargs) [ 627.261177] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.261177] env[63418]: raise e [ 627.261177] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.261177] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 627.261177] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.261177] env[63418]: created_port_ids = self._update_ports_for_instance( [ 627.261177] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.261177] env[63418]: with excutils.save_and_reraise_exception(): [ 627.261177] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.261177] env[63418]: self.force_reraise() [ 627.261177] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.261177] env[63418]: raise self.value [ 627.261177] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.261177] env[63418]: updated_port = self._update_port( [ 627.261177] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.261177] env[63418]: _ensure_no_port_binding_failure(port) [ 627.261177] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.261177] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.261962] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 627.261962] env[63418]: Removing descriptor: 15 [ 627.261962] env[63418]: ERROR nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Traceback (most recent call last): [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] yield resources [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self.driver.spawn(context, instance, image_meta, [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.261962] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] vm_ref = self.build_virtual_machine(instance, [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] for vif in network_info: [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return self._sync_wrapper(fn, *args, **kwargs) [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self.wait() [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self[:] = self._gt.wait() [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return self._exit_event.wait() [ 627.262326] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] result = hub.switch() [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return self.greenlet.switch() [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] result = function(*args, **kwargs) [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return func(*args, **kwargs) [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] raise e [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] nwinfo = self.network_api.allocate_for_instance( [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.262674] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] created_port_ids = self._update_ports_for_instance( [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] with excutils.save_and_reraise_exception(): [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self.force_reraise() [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] raise self.value [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] updated_port = self._update_port( [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] _ensure_no_port_binding_failure(port) [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.263055] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] raise exception.PortBindingFailed(port_id=port['id']) [ 627.263410] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 627.263410] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] [ 627.263410] env[63418]: INFO nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Terminating instance [ 627.497272] env[63418]: DEBUG nova.network.neutron [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.545878] env[63418]: DEBUG nova.network.neutron [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.576896] env[63418]: DEBUG nova.compute.utils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.581451] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 627.582055] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.629482] env[63418]: DEBUG nova.policy [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f86180c9f5c24648b36a6c4e2fab5664', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edbddc12bf554662818bd91b41664379', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.771094] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.953600] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Successfully created port: 4c135418-0ece-4b51-a938-195816f01280 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.049814] env[63418]: DEBUG oslo_concurrency.lockutils [req-1b2f89a9-3c5c-42ce-98a1-bce9a57f9c88 req-36ae2142-9399-4b73-a065-293ff43ed96a service nova] Releasing lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.050975] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquired lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.050975] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.087978] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 628.141739] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c44978a-a068-4628-9f13-fefed3faa6db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.150988] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34a3780-b65a-4e58-a92f-9672a17fc98a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.186776] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54467f93-3ac9-493a-92ba-5cf03d326842 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.194435] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f3bd27-9f71-43f4-ab9d-ae77c24349f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.210344] env[63418]: DEBUG nova.compute.provider_tree [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.590693] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.694205] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.716161] env[63418]: DEBUG nova.scheduler.client.report [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 628.909305] env[63418]: ERROR nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 628.909305] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.909305] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.909305] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.909305] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.909305] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.909305] env[63418]: ERROR nova.compute.manager raise self.value [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.909305] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.909305] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.909305] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.909724] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.909724] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.909724] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 628.909724] env[63418]: ERROR nova.compute.manager [ 628.909724] env[63418]: Traceback (most recent call last): [ 628.909724] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.909724] env[63418]: listener.cb(fileno) [ 628.909724] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.909724] env[63418]: result = function(*args, **kwargs) [ 628.909724] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.909724] env[63418]: return func(*args, **kwargs) [ 628.909724] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 628.909724] env[63418]: raise e [ 628.909724] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.909724] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 628.909724] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.909724] env[63418]: created_port_ids = self._update_ports_for_instance( [ 628.909724] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.909724] env[63418]: with excutils.save_and_reraise_exception(): [ 628.909724] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.909724] env[63418]: self.force_reraise() [ 628.909724] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.909724] env[63418]: raise self.value [ 628.909724] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.909724] env[63418]: updated_port = self._update_port( [ 628.909724] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.909724] env[63418]: _ensure_no_port_binding_failure(port) [ 628.909724] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.909724] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.910501] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 628.910501] env[63418]: Removing descriptor: 15 [ 629.106654] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 629.132421] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.132675] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.132830] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.133010] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.134291] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.134291] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.134291] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.134291] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.134682] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.134913] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.135236] env[63418]: DEBUG nova.virt.hardware [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.136416] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ac8620-e716-4a4c-841d-395f2332f829 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.146546] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56eafba-b4fa-412e-9bd8-ba150984de43 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.166461] env[63418]: ERROR nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Traceback (most recent call last): [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] yield resources [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self.driver.spawn(context, instance, image_meta, [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] vm_ref = self.build_virtual_machine(instance, [ 629.166461] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] for vif in network_info: [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] return self._sync_wrapper(fn, *args, **kwargs) [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self.wait() [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self[:] = self._gt.wait() [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] return self._exit_event.wait() [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.166847] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] current.throw(*self._exc) [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] result = function(*args, **kwargs) [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] return func(*args, **kwargs) [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] raise e [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] nwinfo = self.network_api.allocate_for_instance( [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] created_port_ids = self._update_ports_for_instance( [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] with excutils.save_and_reraise_exception(): [ 629.167250] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self.force_reraise() [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] raise self.value [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] updated_port = self._update_port( [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] _ensure_no_port_binding_failure(port) [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] raise exception.PortBindingFailed(port_id=port['id']) [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 629.167586] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] [ 629.167586] env[63418]: INFO nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Terminating instance [ 629.198181] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Releasing lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.198604] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 629.198789] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.199078] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea60c9e5-1b47-44c0-8b9c-6efc3fde4949 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.208199] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aab1854-8614-4328-ba6f-b76cf24c27d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.221532] env[63418]: DEBUG nova.compute.manager [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Received event network-vif-deleted-04d277c1-ffc4-4d28-bff1-4287ba200126 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 629.221773] env[63418]: DEBUG nova.compute.manager [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Received event network-changed-4c135418-0ece-4b51-a938-195816f01280 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 629.221967] env[63418]: DEBUG nova.compute.manager [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Refreshing instance network info cache due to event network-changed-4c135418-0ece-4b51-a938-195816f01280. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 629.222250] env[63418]: DEBUG oslo_concurrency.lockutils [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] Acquiring lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.222539] env[63418]: DEBUG oslo_concurrency.lockutils [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] Acquired lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.222658] env[63418]: DEBUG nova.network.neutron [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Refreshing network info cache for port 4c135418-0ece-4b51-a938-195816f01280 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 629.227016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.155s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.227016] env[63418]: ERROR nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Traceback (most recent call last): [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self.driver.spawn(context, instance, image_meta, [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.227016] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] vm_ref = self.build_virtual_machine(instance, [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] for vif in network_info: [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return self._sync_wrapper(fn, *args, **kwargs) [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self.wait() [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self[:] = self._gt.wait() [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return self._exit_event.wait() [ 629.227938] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] result = hub.switch() [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return self.greenlet.switch() [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] result = function(*args, **kwargs) [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] return func(*args, **kwargs) [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] raise e [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] nwinfo = self.network_api.allocate_for_instance( [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.228308] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] created_port_ids = self._update_ports_for_instance( [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] with excutils.save_and_reraise_exception(): [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] self.force_reraise() [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] raise self.value [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] updated_port = self._update_port( [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] _ensure_no_port_binding_failure(port) [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.228655] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] raise exception.PortBindingFailed(port_id=port['id']) [ 629.228969] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] nova.exception.PortBindingFailed: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. [ 629.228969] env[63418]: ERROR nova.compute.manager [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] [ 629.228969] env[63418]: DEBUG nova.compute.utils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.228969] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 24.186s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.228969] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.228969] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 629.229180] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.109s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.229854] env[63418]: INFO nova.compute.claims [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.237238] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Build of instance 9b2d0dfa-6820-451f-b86d-590f89928278 was re-scheduled: Binding failed for port d6b6de70-fcba-4f83-8d0b-58583df82063, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 629.237878] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 629.238159] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquiring lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.238344] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Acquired lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.238526] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.240846] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1932a97-b16e-4ffc-8ae8-2239caf0528c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.243513] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 584a7bc1-4651-49ff-a03c-94159f29f30f could not be found. [ 629.243786] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.243998] env[63418]: INFO nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 629.244297] env[63418]: DEBUG oslo.service.loopingcall [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.246017] env[63418]: DEBUG nova.compute.manager [-] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 629.246017] env[63418]: DEBUG nova.network.neutron [-] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.252816] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6394f805-933b-4dcc-9ac6-b0f0a461ed3c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.270175] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52fdae8-9935-4ca1-9c25-f426b79f8285 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.273591] env[63418]: DEBUG nova.network.neutron [-] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.279363] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c1f85b-077a-463b-9e98-3ec70b6c29eb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.312008] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181538MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 629.312233] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.672757] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.750198] env[63418]: DEBUG nova.network.neutron [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.766036] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.775512] env[63418]: DEBUG nova.network.neutron [-] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.831605] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.840182] env[63418]: DEBUG nova.network.neutron [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.985439] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.985638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.279903] env[63418]: INFO nova.compute.manager [-] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Took 1.03 seconds to deallocate network for instance. [ 630.283531] env[63418]: DEBUG nova.compute.claims [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.283653] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.335750] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Releasing lock "refresh_cache-9b2d0dfa-6820-451f-b86d-590f89928278" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.335974] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 630.336146] env[63418]: DEBUG nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.336897] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.343860] env[63418]: DEBUG oslo_concurrency.lockutils [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] Releasing lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.343860] env[63418]: DEBUG nova.compute.manager [req-b2f6d939-e88a-4bdf-b9d1-1bac91a2b1d3 req-e65f7136-b44c-4d7b-a8a7-baa440b88d8b service nova] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Received event network-vif-deleted-4c135418-0ece-4b51-a938-195816f01280 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 630.344456] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquired lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.344456] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.356783] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.750384] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b4c7af-d3b2-471a-8ca4-295496a9ee6f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.758605] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e2b04a-9c5d-4d29-914c-a7d1524f7509 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.789035] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adc7373-9ed1-4158-8f4f-538536db6f9f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.797170] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce945a4b-81b5-4064-8db2-59d4988af3ad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.812362] env[63418]: DEBUG nova.compute.provider_tree [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.860276] env[63418]: DEBUG nova.network.neutron [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.865608] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.896744] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.897007] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.947521] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.315194] env[63418]: DEBUG nova.scheduler.client.report [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 631.363471] env[63418]: INFO nova.compute.manager [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] [instance: 9b2d0dfa-6820-451f-b86d-590f89928278] Took 1.03 seconds to deallocate network for instance. [ 631.450370] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Releasing lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.450818] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 631.451021] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.451337] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd6dc762-6198-43f0-9033-dccdf3a535a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.461792] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab62a24-8780-47f9-b6ea-e89896cfab2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.487651] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 938304f0-1b30-4848-a82d-eb5acf732677 could not be found. [ 631.487911] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.488152] env[63418]: INFO nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Took 0.04 seconds to destroy the instance on the hypervisor. [ 631.488360] env[63418]: DEBUG oslo.service.loopingcall [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.488597] env[63418]: DEBUG nova.compute.manager [-] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 631.488714] env[63418]: DEBUG nova.network.neutron [-] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.507737] env[63418]: DEBUG nova.network.neutron [-] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.820802] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.821754] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 631.825088] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.202s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.826737] env[63418]: INFO nova.compute.claims [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.011154] env[63418]: DEBUG nova.network.neutron [-] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.334043] env[63418]: DEBUG nova.compute.utils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.335758] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 632.335931] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 632.397880] env[63418]: DEBUG nova.policy [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa9ae41d400c495a8b6142d6687b8197', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f3d44332bde412693353e5a0b0ea44d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.429464] env[63418]: INFO nova.scheduler.client.report [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Deleted allocations for instance 9b2d0dfa-6820-451f-b86d-590f89928278 [ 632.513604] env[63418]: INFO nova.compute.manager [-] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Took 1.02 seconds to deallocate network for instance. [ 632.519980] env[63418]: DEBUG nova.compute.claims [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.520193] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.840952] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 632.853734] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Successfully created port: 40bca8e6-54c6-45ea-b1b3-f42428262085 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.940893] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f5bcdba-8b7c-443d-81e8-fe12e2a33f98 tempest-ServerDiagnosticsNegativeTest-550514265 tempest-ServerDiagnosticsNegativeTest-550514265-project-member] Lock "9b2d0dfa-6820-451f-b86d-590f89928278" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.421s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.372756] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2052a1d-780e-4bd5-b595-246b85da222d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.381159] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba39e3c-3ecf-45d3-ae91-c06221b826ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.415012] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ffd0ef-3424-4927-874a-ff7461eb3347 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.424858] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c683373-7dc1-44ba-bb2e-f294c7051713 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.441012] env[63418]: DEBUG nova.compute.provider_tree [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.448114] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 633.856748] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 633.886674] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.886907] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.887077] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.887250] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.887393] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.889512] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.889512] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.889512] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.889512] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.889512] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.889721] env[63418]: DEBUG nova.virt.hardware [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.890960] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ceba8f8-fc91-411c-bb8a-36eb58bb228d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.901631] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31fa4a1-5a76-4456-9a74-c9f8f448f813 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.945087] env[63418]: DEBUG nova.scheduler.client.report [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 633.981351] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.376862] env[63418]: DEBUG nova.compute.manager [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Received event network-changed-40bca8e6-54c6-45ea-b1b3-f42428262085 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 634.377074] env[63418]: DEBUG nova.compute.manager [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Refreshing instance network info cache due to event network-changed-40bca8e6-54c6-45ea-b1b3-f42428262085. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 634.377290] env[63418]: DEBUG oslo_concurrency.lockutils [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] Acquiring lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.377462] env[63418]: DEBUG oslo_concurrency.lockutils [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] Acquired lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.377627] env[63418]: DEBUG nova.network.neutron [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Refreshing network info cache for port 40bca8e6-54c6-45ea-b1b3-f42428262085 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.455417] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.455996] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 634.459378] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.793s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.460804] env[63418]: INFO nova.compute.claims [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.771432] env[63418]: ERROR nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 634.771432] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.771432] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.771432] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.771432] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.771432] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.771432] env[63418]: ERROR nova.compute.manager raise self.value [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.771432] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.771432] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.771432] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.771974] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.771974] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.771974] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 634.771974] env[63418]: ERROR nova.compute.manager [ 634.771974] env[63418]: Traceback (most recent call last): [ 634.771974] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.771974] env[63418]: listener.cb(fileno) [ 634.771974] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.771974] env[63418]: result = function(*args, **kwargs) [ 634.771974] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.771974] env[63418]: return func(*args, **kwargs) [ 634.771974] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 634.771974] env[63418]: raise e [ 634.771974] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.771974] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 634.771974] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.771974] env[63418]: created_port_ids = self._update_ports_for_instance( [ 634.771974] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.771974] env[63418]: with excutils.save_and_reraise_exception(): [ 634.771974] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.771974] env[63418]: self.force_reraise() [ 634.771974] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.771974] env[63418]: raise self.value [ 634.771974] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.771974] env[63418]: updated_port = self._update_port( [ 634.771974] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.771974] env[63418]: _ensure_no_port_binding_failure(port) [ 634.771974] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.771974] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.772746] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 634.772746] env[63418]: Removing descriptor: 15 [ 634.772746] env[63418]: ERROR nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Traceback (most recent call last): [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] yield resources [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self.driver.spawn(context, instance, image_meta, [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.772746] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] vm_ref = self.build_virtual_machine(instance, [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] for vif in network_info: [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return self._sync_wrapper(fn, *args, **kwargs) [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self.wait() [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self[:] = self._gt.wait() [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return self._exit_event.wait() [ 634.773092] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] result = hub.switch() [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return self.greenlet.switch() [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] result = function(*args, **kwargs) [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return func(*args, **kwargs) [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] raise e [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] nwinfo = self.network_api.allocate_for_instance( [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.773475] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] created_port_ids = self._update_ports_for_instance( [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] with excutils.save_and_reraise_exception(): [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self.force_reraise() [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] raise self.value [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] updated_port = self._update_port( [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] _ensure_no_port_binding_failure(port) [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.773824] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] raise exception.PortBindingFailed(port_id=port['id']) [ 634.774146] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 634.774146] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] [ 634.774146] env[63418]: INFO nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Terminating instance [ 634.899949] env[63418]: DEBUG nova.network.neutron [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.972020] env[63418]: DEBUG nova.compute.utils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.976475] env[63418]: DEBUG nova.network.neutron [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.977993] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 634.978182] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.023742] env[63418]: DEBUG nova.policy [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f86180c9f5c24648b36a6c4e2fab5664', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edbddc12bf554662818bd91b41664379', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.275495] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.285847] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Successfully created port: 961af346-5a23-4010-9158-e5f874e5240a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.478624] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 635.495243] env[63418]: DEBUG oslo_concurrency.lockutils [req-9af43ee5-1fbe-425e-87a2-7ac890b87d69 req-24a41de5-ac5e-4653-a458-33b2bac9a318 service nova] Releasing lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.495522] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquired lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.495738] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.035763] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.080351] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cc0923-c785-4c98-8923-4c5f1e649daf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.088765] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5051aa-441c-4dd2-89c0-13ab44189803 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.121692] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6c4fe6-08e3-4448-a08d-83ac6824757a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.129685] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad03399-a1cc-43ac-8a56-54e6d61dadb9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.145225] env[63418]: DEBUG nova.compute.provider_tree [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.188042] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.205279] env[63418]: ERROR nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 636.205279] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.205279] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.205279] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.205279] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.205279] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.205279] env[63418]: ERROR nova.compute.manager raise self.value [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.205279] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.205279] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.205279] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.205733] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.205733] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.205733] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 636.205733] env[63418]: ERROR nova.compute.manager [ 636.205733] env[63418]: Traceback (most recent call last): [ 636.205733] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.205733] env[63418]: listener.cb(fileno) [ 636.205733] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.205733] env[63418]: result = function(*args, **kwargs) [ 636.205733] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.205733] env[63418]: return func(*args, **kwargs) [ 636.205733] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 636.205733] env[63418]: raise e [ 636.205733] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.205733] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 636.205733] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.205733] env[63418]: created_port_ids = self._update_ports_for_instance( [ 636.205733] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.205733] env[63418]: with excutils.save_and_reraise_exception(): [ 636.205733] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.205733] env[63418]: self.force_reraise() [ 636.205733] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.205733] env[63418]: raise self.value [ 636.205733] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.205733] env[63418]: updated_port = self._update_port( [ 636.205733] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.205733] env[63418]: _ensure_no_port_binding_failure(port) [ 636.205733] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.205733] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.206435] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 636.206435] env[63418]: Removing descriptor: 15 [ 636.436636] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "18327b62-d022-45e5-9ab0-71791491e61f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.436869] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "18327b62-d022-45e5-9ab0-71791491e61f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.487605] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 636.523582] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.523825] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.523978] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.524174] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.524316] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.524459] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.524663] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.524817] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.524983] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.525594] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.525796] env[63418]: DEBUG nova.virt.hardware [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.527134] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b565ef-416a-4719-9856-59d87298dc9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.538400] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f0d2bf-518d-4d07-9a3c-160e585a2f33 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.556251] env[63418]: ERROR nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Traceback (most recent call last): [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] yield resources [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self.driver.spawn(context, instance, image_meta, [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] vm_ref = self.build_virtual_machine(instance, [ 636.556251] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] for vif in network_info: [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] return self._sync_wrapper(fn, *args, **kwargs) [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self.wait() [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self[:] = self._gt.wait() [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] return self._exit_event.wait() [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.556620] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] current.throw(*self._exc) [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] result = function(*args, **kwargs) [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] return func(*args, **kwargs) [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] raise e [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] nwinfo = self.network_api.allocate_for_instance( [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] created_port_ids = self._update_ports_for_instance( [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] with excutils.save_and_reraise_exception(): [ 636.556968] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self.force_reraise() [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] raise self.value [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] updated_port = self._update_port( [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] _ensure_no_port_binding_failure(port) [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] raise exception.PortBindingFailed(port_id=port['id']) [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 636.557343] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] [ 636.557343] env[63418]: INFO nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Terminating instance [ 636.612147] env[63418]: DEBUG nova.compute.manager [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Received event network-vif-deleted-40bca8e6-54c6-45ea-b1b3-f42428262085 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 636.612147] env[63418]: DEBUG nova.compute.manager [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Received event network-changed-961af346-5a23-4010-9158-e5f874e5240a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 636.612147] env[63418]: DEBUG nova.compute.manager [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Refreshing instance network info cache due to event network-changed-961af346-5a23-4010-9158-e5f874e5240a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 636.612147] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] Acquiring lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.612147] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] Acquired lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.612589] env[63418]: DEBUG nova.network.neutron [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Refreshing network info cache for port 961af346-5a23-4010-9158-e5f874e5240a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 636.650215] env[63418]: DEBUG nova.scheduler.client.report [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 636.688229] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Releasing lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.688764] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 636.688975] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.689284] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac7518cd-e419-44b5-83cc-9bcb8cc09ed7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.699371] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f850d03-c803-4210-a17b-2be59e71e102 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.725777] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 76e02181-1ccd-43f9-9aaa-cf1918691705 could not be found. [ 636.726035] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.726226] env[63418]: INFO nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Took 0.04 seconds to destroy the instance on the hypervisor. [ 636.726487] env[63418]: DEBUG oslo.service.loopingcall [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.727479] env[63418]: DEBUG nova.compute.manager [-] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 636.727479] env[63418]: DEBUG nova.network.neutron [-] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.755297] env[63418]: DEBUG nova.network.neutron [-] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.062726] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.141356] env[63418]: DEBUG nova.network.neutron [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.160114] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.701s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.160683] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 637.163451] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.393s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.164907] env[63418]: INFO nova.compute.claims [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.249747] env[63418]: DEBUG nova.network.neutron [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.257475] env[63418]: DEBUG nova.network.neutron [-] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.669610] env[63418]: DEBUG nova.compute.utils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.673071] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 637.673267] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 637.730823] env[63418]: DEBUG nova.policy [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c46614205c3d4cf6acfa7176be902236', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a7b4865a2dc45a6803df2987d9b6669', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.912705] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] Releasing lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.912705] env[63418]: DEBUG nova.compute.manager [req-f4234a4f-256e-4d86-84bc-ff6007958344 req-1c2c2c87-34f3-4aa5-a933-cc832b670d0e service nova] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Received event network-vif-deleted-961af346-5a23-4010-9158-e5f874e5240a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 637.912705] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquired lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.912705] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.912705] env[63418]: INFO nova.compute.manager [-] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Took 1.03 seconds to deallocate network for instance. [ 637.912705] env[63418]: DEBUG nova.compute.claims [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.913143] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.173829] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 638.200988] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Successfully created port: cb30e2b7-0ce4-43cc-ac32-5782923ebd6e {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.290250] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.434828] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.594997] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.595269] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.771902] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d487d7-9167-46c2-9d37-812f164a8685 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.782306] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f3bf68-b23d-4390-8e20-d8f7a40d996a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.815719] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ca19f1-2bb8-4c20-9fc5-4d73d7a92ffb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.825716] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4350ff1d-061a-42c1-9493-2477b114fd67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.842859] env[63418]: DEBUG nova.compute.provider_tree [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.938674] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Releasing lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.939020] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 638.939238] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.939595] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d2db10f-729c-45b1-b276-e74e87c558d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.944625] env[63418]: DEBUG nova.compute.manager [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Received event network-changed-cb30e2b7-0ce4-43cc-ac32-5782923ebd6e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 638.944954] env[63418]: DEBUG nova.compute.manager [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Refreshing instance network info cache due to event network-changed-cb30e2b7-0ce4-43cc-ac32-5782923ebd6e. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 638.945040] env[63418]: DEBUG oslo_concurrency.lockutils [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] Acquiring lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.945151] env[63418]: DEBUG oslo_concurrency.lockutils [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] Acquired lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.945276] env[63418]: DEBUG nova.network.neutron [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Refreshing network info cache for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.954086] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e2cb33-25d8-4329-8bf3-75fc47e72d98 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.981390] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b56238f3-3507-49e0-8630-5f5a093c1101 could not be found. [ 638.981639] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.981764] env[63418]: INFO nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Took 0.04 seconds to destroy the instance on the hypervisor. [ 638.982012] env[63418]: DEBUG oslo.service.loopingcall [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.982714] env[63418]: DEBUG nova.compute.manager [-] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 638.982812] env[63418]: DEBUG nova.network.neutron [-] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.997522] env[63418]: DEBUG nova.network.neutron [-] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.144113] env[63418]: ERROR nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 639.144113] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 639.144113] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.144113] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.144113] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.144113] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.144113] env[63418]: ERROR nova.compute.manager raise self.value [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.144113] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.144113] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.144113] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.144644] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.144644] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.144644] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 639.144644] env[63418]: ERROR nova.compute.manager [ 639.144644] env[63418]: Traceback (most recent call last): [ 639.144644] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.144644] env[63418]: listener.cb(fileno) [ 639.144644] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.144644] env[63418]: result = function(*args, **kwargs) [ 639.144644] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.144644] env[63418]: return func(*args, **kwargs) [ 639.144644] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 639.144644] env[63418]: raise e [ 639.144644] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 639.144644] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 639.144644] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.144644] env[63418]: created_port_ids = self._update_ports_for_instance( [ 639.144644] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.144644] env[63418]: with excutils.save_and_reraise_exception(): [ 639.144644] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.144644] env[63418]: self.force_reraise() [ 639.144644] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.144644] env[63418]: raise self.value [ 639.144644] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.144644] env[63418]: updated_port = self._update_port( [ 639.144644] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.144644] env[63418]: _ensure_no_port_binding_failure(port) [ 639.144644] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.144644] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.145331] env[63418]: nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 639.145331] env[63418]: Removing descriptor: 15 [ 639.191847] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 639.225656] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.225797] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.225845] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.228718] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.228718] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.228718] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.228718] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.228718] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.228914] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.228914] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.228914] env[63418]: DEBUG nova.virt.hardware [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.229049] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3788e84-64c0-4c50-8936-49ebdd368ad1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.240842] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a817e9a6-a8fd-4b85-a441-675929802f76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.257118] env[63418]: ERROR nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Traceback (most recent call last): [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] yield resources [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self.driver.spawn(context, instance, image_meta, [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] vm_ref = self.build_virtual_machine(instance, [ 639.257118] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] for vif in network_info: [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] return self._sync_wrapper(fn, *args, **kwargs) [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self.wait() [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self[:] = self._gt.wait() [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] return self._exit_event.wait() [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.257449] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] current.throw(*self._exc) [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] result = function(*args, **kwargs) [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] return func(*args, **kwargs) [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] raise e [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] nwinfo = self.network_api.allocate_for_instance( [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] created_port_ids = self._update_ports_for_instance( [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] with excutils.save_and_reraise_exception(): [ 639.257756] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self.force_reraise() [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] raise self.value [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] updated_port = self._update_port( [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] _ensure_no_port_binding_failure(port) [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] raise exception.PortBindingFailed(port_id=port['id']) [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 639.258075] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] [ 639.258075] env[63418]: INFO nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Terminating instance [ 639.346074] env[63418]: DEBUG nova.scheduler.client.report [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 639.465325] env[63418]: DEBUG nova.network.neutron [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.500282] env[63418]: DEBUG nova.network.neutron [-] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.556402] env[63418]: DEBUG nova.network.neutron [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.762568] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquiring lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.851880] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.688s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.852439] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 639.855056] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.025s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.002822] env[63418]: INFO nova.compute.manager [-] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Took 1.02 seconds to deallocate network for instance. [ 640.005300] env[63418]: DEBUG nova.compute.claims [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.005560] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.058722] env[63418]: DEBUG oslo_concurrency.lockutils [req-c15c4ff9-2ebe-4264-8072-99beb3287e28 req-28f1fd4b-7f80-4fc9-a497-8f7b1df8cc85 service nova] Releasing lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.059199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquired lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.059423] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.362775] env[63418]: DEBUG nova.compute.utils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.364299] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 640.364530] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.405482] env[63418]: DEBUG nova.policy [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e3d627138f4b779d5d01ef63bfb829', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55d17b7c317d4aeead84e544d5251363', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.581844] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.665847] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Successfully created port: ade83678-a78b-4563-b665-2ec39a8ab330 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.815224] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e9dd01-b29d-4fa4-b023-686536df0e2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.823654] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9cac5d-6a39-46d7-b7e8-96bb31254617 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.855415] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a55751-d19d-4ed8-9219-69842d4b043b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.859799] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.865667] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b924b0c-e27a-442c-99f4-24e0527858ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.870593] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 640.884330] env[63418]: DEBUG nova.compute.provider_tree [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.019652] env[63418]: DEBUG nova.compute.manager [req-96d0d1b8-027f-4385-b14d-a894bf754a5e req-1623fa80-554f-4dd7-9633-0b9c116ba573 service nova] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Received event network-vif-deleted-cb30e2b7-0ce4-43cc-ac32-5782923ebd6e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 641.368024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Releasing lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.368024] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 641.368024] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.368024] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc07313e-e801-485a-a3c4-c626c656105b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.383192] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a949e0-67e7-402f-b772-4242242f673e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.398078] env[63418]: DEBUG nova.scheduler.client.report [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 641.417539] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59442c77-234c-48c2-872d-85ed9bd8cb80 could not be found. [ 641.417863] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.418106] env[63418]: INFO nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Took 0.05 seconds to destroy the instance on the hypervisor. [ 641.418389] env[63418]: DEBUG oslo.service.loopingcall [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.419330] env[63418]: DEBUG nova.compute.manager [-] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.419330] env[63418]: DEBUG nova.network.neutron [-] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.450071] env[63418]: DEBUG nova.network.neutron [-] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.693198] env[63418]: ERROR nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 641.693198] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.693198] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.693198] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.693198] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.693198] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.693198] env[63418]: ERROR nova.compute.manager raise self.value [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.693198] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.693198] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.693198] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.693740] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.693740] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.693740] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 641.693740] env[63418]: ERROR nova.compute.manager [ 641.693740] env[63418]: Traceback (most recent call last): [ 641.693740] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.693740] env[63418]: listener.cb(fileno) [ 641.693740] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.693740] env[63418]: result = function(*args, **kwargs) [ 641.693740] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.693740] env[63418]: return func(*args, **kwargs) [ 641.693740] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 641.693740] env[63418]: raise e [ 641.693740] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.693740] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 641.693740] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.693740] env[63418]: created_port_ids = self._update_ports_for_instance( [ 641.693740] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.693740] env[63418]: with excutils.save_and_reraise_exception(): [ 641.693740] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.693740] env[63418]: self.force_reraise() [ 641.693740] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.693740] env[63418]: raise self.value [ 641.693740] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.693740] env[63418]: updated_port = self._update_port( [ 641.693740] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.693740] env[63418]: _ensure_no_port_binding_failure(port) [ 641.693740] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.693740] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.694697] env[63418]: nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 641.694697] env[63418]: Removing descriptor: 15 [ 641.880734] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 641.903981] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.049s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.904473] env[63418]: ERROR nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Traceback (most recent call last): [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self.driver.spawn(context, instance, image_meta, [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] vm_ref = self.build_virtual_machine(instance, [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.904473] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] for vif in network_info: [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] return self._sync_wrapper(fn, *args, **kwargs) [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self.wait() [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self[:] = self._gt.wait() [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] return self._exit_event.wait() [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] current.throw(*self._exc) [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.904815] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] result = function(*args, **kwargs) [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] return func(*args, **kwargs) [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] raise e [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] nwinfo = self.network_api.allocate_for_instance( [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] created_port_ids = self._update_ports_for_instance( [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] with excutils.save_and_reraise_exception(): [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] self.force_reraise() [ 641.905200] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] raise self.value [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] updated_port = self._update_port( [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] _ensure_no_port_binding_failure(port) [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] raise exception.PortBindingFailed(port_id=port['id']) [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] nova.exception.PortBindingFailed: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. [ 641.905538] env[63418]: ERROR nova.compute.manager [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] [ 641.905538] env[63418]: DEBUG nova.compute.utils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.908648] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.908867] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.909032] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.909225] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.909373] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.909517] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.909717] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.909871] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.910058] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.910227] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.910395] env[63418]: DEBUG nova.virt.hardware [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.910695] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.447s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.912203] env[63418]: INFO nova.compute.claims [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.915287] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd3dfc0-2a3c-4a29-9210-e8df211a7f2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.918288] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Build of instance 0c55e7b9-cb53-4525-8524-4ef62a51a89f was re-scheduled: Binding failed for port 6272afaa-e1e6-4386-82ea-26eb9aba171e, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 641.918723] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 641.918945] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquiring lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.919105] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Acquired lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.919265] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.927462] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92821d00-b005-4579-a998-756a13e8edd7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.945672] env[63418]: ERROR nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] Traceback (most recent call last): [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] yield resources [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self.driver.spawn(context, instance, image_meta, [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] vm_ref = self.build_virtual_machine(instance, [ 641.945672] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] for vif in network_info: [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] return self._sync_wrapper(fn, *args, **kwargs) [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self.wait() [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self[:] = self._gt.wait() [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] return self._exit_event.wait() [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.946050] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] current.throw(*self._exc) [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] result = function(*args, **kwargs) [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] return func(*args, **kwargs) [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] raise e [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] nwinfo = self.network_api.allocate_for_instance( [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] created_port_ids = self._update_ports_for_instance( [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] with excutils.save_and_reraise_exception(): [ 641.946379] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self.force_reraise() [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] raise self.value [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] updated_port = self._update_port( [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] _ensure_no_port_binding_failure(port) [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] raise exception.PortBindingFailed(port_id=port['id']) [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 641.946679] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] [ 641.946679] env[63418]: INFO nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Terminating instance [ 641.952067] env[63418]: DEBUG nova.network.neutron [-] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.436763] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.450567] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquiring lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.450854] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquired lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.450960] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.453759] env[63418]: INFO nova.compute.manager [-] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Took 1.03 seconds to deallocate network for instance. [ 642.456833] env[63418]: DEBUG nova.compute.claims [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.457543] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.522194] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.969275] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.025524] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Releasing lock "refresh_cache-0c55e7b9-cb53-4525-8524-4ef62a51a89f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.025906] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 643.025906] env[63418]: DEBUG nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 643.026077] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.041666] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.044559] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.089937] env[63418]: DEBUG nova.compute.manager [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] [instance: b24d095a-6814-413b-bf89-511e22df740f] Received event network-changed-ade83678-a78b-4563-b665-2ec39a8ab330 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 643.090155] env[63418]: DEBUG nova.compute.manager [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] [instance: b24d095a-6814-413b-bf89-511e22df740f] Refreshing instance network info cache due to event network-changed-ade83678-a78b-4563-b665-2ec39a8ab330. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 643.090344] env[63418]: DEBUG oslo_concurrency.lockutils [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] Acquiring lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.348752] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a87e34a-bc08-4b17-927a-059d7e8a348c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.358026] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f134a49-154a-4d62-80b3-021ed0d3a8d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.390437] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d039b2-ba81-4a6e-b290-dd890d1fae1e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.398397] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e714da-c111-471d-9725-85c494c33181 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.412033] env[63418]: DEBUG nova.compute.provider_tree [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.544639] env[63418]: DEBUG nova.network.neutron [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.548470] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Releasing lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.548470] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 643.549433] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.549433] env[63418]: DEBUG oslo_concurrency.lockutils [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] Acquired lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.549433] env[63418]: DEBUG nova.network.neutron [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] [instance: b24d095a-6814-413b-bf89-511e22df740f] Refreshing network info cache for port ade83678-a78b-4563-b665-2ec39a8ab330 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 643.550092] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f7d90e7-1a52-4f4b-b035-7c07938340f7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.560820] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c853030-e412-4734-bebb-f6be468efaa7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.587433] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b24d095a-6814-413b-bf89-511e22df740f could not be found. [ 643.587660] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.587835] env[63418]: INFO nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 643.588080] env[63418]: DEBUG oslo.service.loopingcall [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.588325] env[63418]: DEBUG nova.compute.manager [-] [instance: b24d095a-6814-413b-bf89-511e22df740f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 643.588391] env[63418]: DEBUG nova.network.neutron [-] [instance: b24d095a-6814-413b-bf89-511e22df740f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.603630] env[63418]: DEBUG nova.network.neutron [-] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.915737] env[63418]: DEBUG nova.scheduler.client.report [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 644.047871] env[63418]: INFO nova.compute.manager [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] [instance: 0c55e7b9-cb53-4525-8524-4ef62a51a89f] Took 1.02 seconds to deallocate network for instance. [ 644.068413] env[63418]: DEBUG nova.network.neutron [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.107034] env[63418]: DEBUG nova.network.neutron [-] [instance: b24d095a-6814-413b-bf89-511e22df740f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.142203] env[63418]: DEBUG nova.network.neutron [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] [instance: b24d095a-6814-413b-bf89-511e22df740f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.570791] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.570791] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 644.570791] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.818s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.570791] env[63418]: INFO nova.compute.claims [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.610606] env[63418]: INFO nova.compute.manager [-] [instance: b24d095a-6814-413b-bf89-511e22df740f] Took 1.02 seconds to deallocate network for instance. [ 644.612083] env[63418]: DEBUG nova.compute.claims [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.612260] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.644868] env[63418]: DEBUG oslo_concurrency.lockutils [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] Releasing lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.644868] env[63418]: DEBUG nova.compute.manager [req-938c21b7-f758-47df-a4dc-0284fd70f22d req-29bc8520-c183-42a6-a7d3-277fc0374cf3 service nova] [instance: b24d095a-6814-413b-bf89-511e22df740f] Received event network-vif-deleted-ade83678-a78b-4563-b665-2ec39a8ab330 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 644.932760] env[63418]: DEBUG nova.compute.utils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.936566] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 644.937627] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 644.975859] env[63418]: DEBUG nova.policy [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43c4e042e60c4db89a9ea3f5f87cf456', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1de059e6836948a09660e85b598cc8bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 645.084445] env[63418]: INFO nova.scheduler.client.report [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Deleted allocations for instance 0c55e7b9-cb53-4525-8524-4ef62a51a89f [ 645.291816] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Successfully created port: d325e415-e7a8-482f-aa6c-4eb998861d38 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.438063] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 645.597406] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2289965e-b560-40c4-8e78-9eb9c415a424 tempest-ServerMetadataNegativeTestJSON-1471309201 tempest-ServerMetadataNegativeTestJSON-1471309201-project-member] Lock "0c55e7b9-cb53-4525-8524-4ef62a51a89f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.169s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.914655] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344b093e-8f7f-49eb-8c9c-24cccd41cd47 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.930172] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0192c6d1-9eb0-4f63-8ff0-b69ee80152e7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.968954] env[63418]: INFO nova.virt.block_device [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Booting with volume f5edfa69-6e8d-4438-9da4-a1bc888c50e0 at /dev/sda [ 645.973626] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b260c56-cb5f-4e75-b411-8ada1df32d8d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.991632] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa39a5a-dc0f-4755-8052-5be80df6481a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.007591] env[63418]: DEBUG nova.compute.provider_tree [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.020571] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a300e96-4e8b-426f-b795-893fc4f1102a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.030203] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d6ab2c-69b8-4c73-a391-f7189df14a9e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.043616] env[63418]: DEBUG nova.compute.manager [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Received event network-changed-d325e415-e7a8-482f-aa6c-4eb998861d38 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 646.043616] env[63418]: DEBUG nova.compute.manager [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Refreshing instance network info cache due to event network-changed-d325e415-e7a8-482f-aa6c-4eb998861d38. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 646.043616] env[63418]: DEBUG oslo_concurrency.lockutils [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] Acquiring lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.043616] env[63418]: DEBUG oslo_concurrency.lockutils [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] Acquired lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.043761] env[63418]: DEBUG nova.network.neutron [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Refreshing network info cache for port d325e415-e7a8-482f-aa6c-4eb998861d38 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.058236] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-323c827b-bf05-4111-9f00-9f3ed4e32377 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.069041] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ae0c93-b66f-4357-bcef-a7d1f2097f73 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.092431] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fe0162-2696-4064-97be-325fa79cd8d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.099600] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425a0440-f5fe-46e8-9823-0567e4af15c9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.104153] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 646.118065] env[63418]: DEBUG nova.virt.block_device [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Updating existing volume attachment record: 479198a6-b942-4577-989f-5775a8590e0d {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 646.195741] env[63418]: ERROR nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 646.195741] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.195741] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.195741] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.195741] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.195741] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.195741] env[63418]: ERROR nova.compute.manager raise self.value [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.195741] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.195741] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.195741] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.196241] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.196241] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.196241] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 646.196241] env[63418]: ERROR nova.compute.manager [ 646.196241] env[63418]: Traceback (most recent call last): [ 646.196241] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.196241] env[63418]: listener.cb(fileno) [ 646.196241] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.196241] env[63418]: result = function(*args, **kwargs) [ 646.196241] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.196241] env[63418]: return func(*args, **kwargs) [ 646.196241] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 646.196241] env[63418]: raise e [ 646.196241] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.196241] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 646.196241] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.196241] env[63418]: created_port_ids = self._update_ports_for_instance( [ 646.196241] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.196241] env[63418]: with excutils.save_and_reraise_exception(): [ 646.196241] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.196241] env[63418]: self.force_reraise() [ 646.196241] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.196241] env[63418]: raise self.value [ 646.196241] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.196241] env[63418]: updated_port = self._update_port( [ 646.196241] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.196241] env[63418]: _ensure_no_port_binding_failure(port) [ 646.196241] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.196241] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.197034] env[63418]: nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 646.197034] env[63418]: Removing descriptor: 15 [ 646.511296] env[63418]: DEBUG nova.scheduler.client.report [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 646.563715] env[63418]: DEBUG nova.network.neutron [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.628167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.638505] env[63418]: DEBUG nova.network.neutron [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.018565] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.019126] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 647.025567] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.713s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.143858] env[63418]: DEBUG oslo_concurrency.lockutils [req-d5b03389-0932-4f6b-943d-f018178c1914 req-c649d816-c409-4263-a22f-a574a349f7ce service nova] Releasing lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.526944] env[63418]: DEBUG nova.compute.utils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.533019] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 647.533019] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.591584] env[63418]: DEBUG nova.policy [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.036389] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 648.066232] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 584a7bc1-4651-49ff-a03c-94159f29f30f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066232] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 938304f0-1b30-4848-a82d-eb5acf732677 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066232] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 76e02181-1ccd-43f9-9aaa-cf1918691705 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066232] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance b56238f3-3507-49e0-8630-5f5a093c1101 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066458] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 59442c77-234c-48c2-872d-85ed9bd8cb80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066458] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance b24d095a-6814-413b-bf89-511e22df740f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066458] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 601f72f4-05bc-4d73-9b57-612cad2d8c62 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.066458] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6d47c52c-c1fb-4ade-a3ba-dc393716eb79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.080918] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Successfully created port: cc684a4f-1b55-4017-8a0f-5787cd0654bc {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.162392] env[63418]: DEBUG nova.compute.manager [req-9fea4408-88be-4d36-9a66-6bad74abff06 req-e0d3c132-ffc5-4a47-813e-47171d8f1ead service nova] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Received event network-vif-deleted-d325e415-e7a8-482f-aa6c-4eb998861d38 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 648.235960] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 648.236549] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.236760] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.236910] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.237099] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.237276] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.237391] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.237690] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.237775] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.238274] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.238274] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.238595] env[63418]: DEBUG nova.virt.hardware [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.239718] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12207fb0-6f8b-4a3b-b29d-6a17b6f5cf7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.250021] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65197610-06fc-4a09-bfb7-b183db1e724c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.266920] env[63418]: ERROR nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Traceback (most recent call last): [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] yield resources [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self.driver.spawn(context, instance, image_meta, [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] vm_ref = self.build_virtual_machine(instance, [ 648.266920] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] for vif in network_info: [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] return self._sync_wrapper(fn, *args, **kwargs) [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self.wait() [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self[:] = self._gt.wait() [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] return self._exit_event.wait() [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 648.267369] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] current.throw(*self._exc) [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] result = function(*args, **kwargs) [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] return func(*args, **kwargs) [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] raise e [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] nwinfo = self.network_api.allocate_for_instance( [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] created_port_ids = self._update_ports_for_instance( [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] with excutils.save_and_reraise_exception(): [ 648.267829] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self.force_reraise() [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] raise self.value [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] updated_port = self._update_port( [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] _ensure_no_port_binding_failure(port) [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] raise exception.PortBindingFailed(port_id=port['id']) [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 648.268230] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] [ 648.268230] env[63418]: INFO nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Terminating instance [ 648.574492] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance c590bbfe-1ab8-4e7a-a3aa-89e1f579d756 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.773675] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquiring lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.773969] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquired lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.774187] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.979417] env[63418]: ERROR nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 648.979417] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 648.979417] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.979417] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.979417] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.979417] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.979417] env[63418]: ERROR nova.compute.manager raise self.value [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.979417] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.979417] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.979417] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.979862] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.979862] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.979862] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 648.979862] env[63418]: ERROR nova.compute.manager [ 648.979862] env[63418]: Traceback (most recent call last): [ 648.979862] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.979862] env[63418]: listener.cb(fileno) [ 648.979862] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.979862] env[63418]: result = function(*args, **kwargs) [ 648.979862] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.979862] env[63418]: return func(*args, **kwargs) [ 648.979862] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 648.979862] env[63418]: raise e [ 648.979862] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 648.979862] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 648.979862] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.979862] env[63418]: created_port_ids = self._update_ports_for_instance( [ 648.979862] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.979862] env[63418]: with excutils.save_and_reraise_exception(): [ 648.979862] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.979862] env[63418]: self.force_reraise() [ 648.979862] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.979862] env[63418]: raise self.value [ 648.979862] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.979862] env[63418]: updated_port = self._update_port( [ 648.979862] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.979862] env[63418]: _ensure_no_port_binding_failure(port) [ 648.979862] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.979862] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.980583] env[63418]: nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 648.980583] env[63418]: Removing descriptor: 15 [ 649.045205] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 649.070391] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.070639] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.070792] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.070968] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.071127] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.071275] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.071476] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.071631] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.071827] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.071994] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.072178] env[63418]: DEBUG nova.virt.hardware [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.073816] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bf2de0-4064-41c3-845c-71e0ad9c381b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.077113] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance f6d62d1b-0333-40ac-92e8-a798e58d495c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.084951] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0c69a5-38cf-4ef8-8feb-d3ac6b5a5369 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.098697] env[63418]: ERROR nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Traceback (most recent call last): [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] yield resources [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self.driver.spawn(context, instance, image_meta, [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] vm_ref = self.build_virtual_machine(instance, [ 649.098697] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] for vif in network_info: [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] return self._sync_wrapper(fn, *args, **kwargs) [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self.wait() [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self[:] = self._gt.wait() [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] return self._exit_event.wait() [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 649.099038] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] current.throw(*self._exc) [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] result = function(*args, **kwargs) [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] return func(*args, **kwargs) [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] raise e [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] nwinfo = self.network_api.allocate_for_instance( [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] created_port_ids = self._update_ports_for_instance( [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] with excutils.save_and_reraise_exception(): [ 649.099364] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self.force_reraise() [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] raise self.value [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] updated_port = self._update_port( [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] _ensure_no_port_binding_failure(port) [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] raise exception.PortBindingFailed(port_id=port['id']) [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 649.099686] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] [ 649.099686] env[63418]: INFO nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Terminating instance [ 649.298197] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.399177] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.580535] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 8e39c8ef-103a-45f4-8630-bfd51b38445d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.603824] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.604054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.604247] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.901756] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Releasing lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.902360] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 649.902692] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d7d00fb-dd97-4ad5-9ae1-38f4eed00bf7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.912542] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9497c4-c211-464f-b433-b79bcd48c1cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.936870] env[63418]: WARNING nova.virt.vmwareapi.driver [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 601f72f4-05bc-4d73-9b57-612cad2d8c62 could not be found. [ 649.937106] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.937388] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d53a127-7b75-4429-9eb7-36b806860edc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.946226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7cd9b5-28dc-4493-8276-a972bada5b0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.975019] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 601f72f4-05bc-4d73-9b57-612cad2d8c62 could not be found. [ 649.975019] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.975019] env[63418]: INFO nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Took 0.07 seconds to destroy the instance on the hypervisor. [ 649.975019] env[63418]: DEBUG oslo.service.loopingcall [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.975019] env[63418]: DEBUG nova.compute.manager [-] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 649.975019] env[63418]: DEBUG nova.network.neutron [-] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.987743] env[63418]: DEBUG nova.network.neutron [-] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.083723] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 69651fab-08a7-4832-8c91-2fa01d186860 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.121903] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.188884] env[63418]: DEBUG nova.compute.manager [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Received event network-changed-cc684a4f-1b55-4017-8a0f-5787cd0654bc {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 650.189090] env[63418]: DEBUG nova.compute.manager [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Refreshing instance network info cache due to event network-changed-cc684a4f-1b55-4017-8a0f-5787cd0654bc. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 650.189281] env[63418]: DEBUG oslo_concurrency.lockutils [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] Acquiring lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.194256] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.490716] env[63418]: DEBUG nova.network.neutron [-] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.586819] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 2edbeef4-be3b-4a1d-8895-8410b71b6b18 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.698496] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.698945] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 650.699150] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.699469] env[63418]: DEBUG oslo_concurrency.lockutils [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] Acquired lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.699641] env[63418]: DEBUG nova.network.neutron [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Refreshing network info cache for port cc684a4f-1b55-4017-8a0f-5787cd0654bc {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.700728] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a71bbb9-140e-464e-b1ba-d5749664f3ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.711587] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f40ef8-f1da-4354-a648-93a9b1625963 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.735811] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6d47c52c-c1fb-4ade-a3ba-dc393716eb79 could not be found. [ 650.735811] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.735811] env[63418]: INFO nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Took 0.04 seconds to destroy the instance on the hypervisor. [ 650.736056] env[63418]: DEBUG oslo.service.loopingcall [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.736261] env[63418]: DEBUG nova.compute.manager [-] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 650.736417] env[63418]: DEBUG nova.network.neutron [-] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.752317] env[63418]: DEBUG nova.network.neutron [-] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.993453] env[63418]: INFO nova.compute.manager [-] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Took 1.02 seconds to deallocate network for instance. [ 651.090042] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 8c78df16-faf1-4007-8492-bdc7cbfc608a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.220962] env[63418]: DEBUG nova.network.neutron [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.254599] env[63418]: DEBUG nova.network.neutron [-] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.302573] env[63418]: DEBUG nova.network.neutron [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.548012] env[63418]: INFO nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Took 0.55 seconds to detach 1 volumes for instance. [ 651.550208] env[63418]: DEBUG nova.compute.claims [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.550456] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.592712] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6a80e0c1-2088-4de3-9032-d05ff48c3fcb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.757032] env[63418]: INFO nova.compute.manager [-] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Took 1.02 seconds to deallocate network for instance. [ 651.759056] env[63418]: DEBUG nova.compute.claims [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.759320] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.805184] env[63418]: DEBUG oslo_concurrency.lockutils [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] Releasing lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.805400] env[63418]: DEBUG nova.compute.manager [req-c7c743f6-acb8-42ab-874e-37d81580fffe req-2376407f-3295-452c-aa93-df1efe65b240 service nova] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Received event network-vif-deleted-cc684a4f-1b55-4017-8a0f-5787cd0654bc {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 652.096899] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 99d00df2-6bf3-4ffe-b77d-f44ada631c8b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.599909] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance c79aa3d1-bbca-45cf-9f29-40a2720c4437 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.104062] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 96183939-654d-4510-adf2-303f41f94640 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.606700] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.109522] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance e7a4d56b-f785-44b4-acb3-6390a4bad83c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.613449] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 702be472-a4bc-4089-94de-4355ee00684a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.116197] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6b81dbe2-aa9e-4561-962b-2af167234b90 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.619493] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 17a055e4-13da-4984-ae64-af1ae9d652f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.123047] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance c0fd1999-346d-4c12-9b8c-ab7e21ec4227 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.626830] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance d76a008c-9bd9-420b-873d-4f7d7f25b8ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.129780] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance c4cb1770-b3c7-4f6b-ba11-85310f2175a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.633885] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.136722] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7615244b-fa19-4dbb-a851-604bb6fc2725 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.639565] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 0c151b0c-c383-4421-b506-b7afa95e2072 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.143117] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance f632d71d-498f-4914-9895-8f37187a295f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.647279] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance e3dd6d5e-2b80-4724-9826-6ebfb8181c3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.150720] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.653694] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.156816] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 18327b62-d022-45e5-9ab0-71791491e61f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.659656] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.659932] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 661.660093] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 662.028820] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0861ce-ae07-4d27-822c-c68c705fe8aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.036384] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dd7594-2051-4397-98f3-26584173f565 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.066289] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4723ff-6a49-4651-a1af-a32b2a552a28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.073334] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0464e8-9298-400c-a7b4-8b2ba418ecc7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.086139] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.589412] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 663.095710] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 663.096019] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 16.070s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.096357] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.813s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.972399] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30afcba2-6c33-4a46-9782-baf9faef6a68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.980123] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfc2053-e653-4879-944f-8ce64e57a410 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.009699] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a29a69-13bd-40a9-a03a-1a153d285e06 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.016938] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1a0a3d-52e9-4edc-b144-9614e4933646 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.030497] env[63418]: DEBUG nova.compute.provider_tree [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.533721] env[63418]: DEBUG nova.scheduler.client.report [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 665.039511] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.039929] env[63418]: ERROR nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Traceback (most recent call last): [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self.driver.spawn(context, instance, image_meta, [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] vm_ref = self.build_virtual_machine(instance, [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.039929] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] for vif in network_info: [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return self._sync_wrapper(fn, *args, **kwargs) [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self.wait() [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self[:] = self._gt.wait() [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return self._exit_event.wait() [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] result = hub.switch() [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.040275] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return self.greenlet.switch() [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] result = function(*args, **kwargs) [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] return func(*args, **kwargs) [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] raise e [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] nwinfo = self.network_api.allocate_for_instance( [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] created_port_ids = self._update_ports_for_instance( [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] with excutils.save_and_reraise_exception(): [ 665.040582] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] self.force_reraise() [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] raise self.value [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] updated_port = self._update_port( [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] _ensure_no_port_binding_failure(port) [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] raise exception.PortBindingFailed(port_id=port['id']) [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] nova.exception.PortBindingFailed: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. [ 665.040894] env[63418]: ERROR nova.compute.manager [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] [ 665.041177] env[63418]: DEBUG nova.compute.utils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.042252] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.522s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.044773] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Build of instance 584a7bc1-4651-49ff-a03c-94159f29f30f was re-scheduled: Binding failed for port 04d277c1-ffc4-4d28-bff1-4287ba200126, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 665.045210] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 665.045438] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.045583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquired lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.045736] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.567675] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.620231] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.976497] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92940aff-9da8-4396-bd8e-ca574748fac8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.984366] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f45f8d-dabd-445a-92d6-d9b4f2d7882a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.016020] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1f8cb4-27a2-440d-bff6-b32fe9b4c141 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.020999] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b199c666-1ffd-4cf5-a57a-aa00a8365708 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.038318] env[63418]: DEBUG nova.compute.provider_tree [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.122730] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Releasing lock "refresh_cache-584a7bc1-4651-49ff-a03c-94159f29f30f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.122977] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 666.123141] env[63418]: DEBUG nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 666.123307] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.139619] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.540893] env[63418]: DEBUG nova.scheduler.client.report [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 666.641844] env[63418]: DEBUG nova.network.neutron [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.046064] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.046727] env[63418]: ERROR nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Traceback (most recent call last): [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self.driver.spawn(context, instance, image_meta, [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] vm_ref = self.build_virtual_machine(instance, [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.046727] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] for vif in network_info: [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] return self._sync_wrapper(fn, *args, **kwargs) [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self.wait() [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self[:] = self._gt.wait() [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] return self._exit_event.wait() [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] current.throw(*self._exc) [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.047131] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] result = function(*args, **kwargs) [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] return func(*args, **kwargs) [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] raise e [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] nwinfo = self.network_api.allocate_for_instance( [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] created_port_ids = self._update_ports_for_instance( [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] with excutils.save_and_reraise_exception(): [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] self.force_reraise() [ 667.047526] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] raise self.value [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] updated_port = self._update_port( [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] _ensure_no_port_binding_failure(port) [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] raise exception.PortBindingFailed(port_id=port['id']) [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] nova.exception.PortBindingFailed: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. [ 667.047938] env[63418]: ERROR nova.compute.manager [instance: 938304f0-1b30-4848-a82d-eb5acf732677] [ 667.047938] env[63418]: DEBUG nova.compute.utils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.048999] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.067s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.050551] env[63418]: INFO nova.compute.claims [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.053207] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Build of instance 938304f0-1b30-4848-a82d-eb5acf732677 was re-scheduled: Binding failed for port 4c135418-0ece-4b51-a938-195816f01280, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 667.053654] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 667.053872] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.054077] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquired lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.054180] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.144716] env[63418]: INFO nova.compute.manager [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 584a7bc1-4651-49ff-a03c-94159f29f30f] Took 1.02 seconds to deallocate network for instance. [ 667.574182] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.647645] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.156022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Releasing lock "refresh_cache-938304f0-1b30-4848-a82d-eb5acf732677" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.156022] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 668.156022] env[63418]: DEBUG nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 668.156022] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.173492] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.176122] env[63418]: INFO nova.scheduler.client.report [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Deleted allocations for instance 584a7bc1-4651-49ff-a03c-94159f29f30f [ 668.488085] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be84f36-12f1-4cc3-963b-024d5c53a539 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.496166] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5a42e4-8b4a-4ab4-be32-28866f412cd0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.528679] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83478b06-6276-4dce-8ad4-a4015fcf853b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.536950] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52116626-1700-49e6-960d-c46ca43f1a3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.550037] env[63418]: DEBUG nova.compute.provider_tree [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.683807] env[63418]: DEBUG nova.network.neutron [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.688126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a1bb4e36-8058-4f1f-aef6-f4b5fb46a1e1 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "584a7bc1-4651-49ff-a03c-94159f29f30f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.733s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.052921] env[63418]: DEBUG nova.scheduler.client.report [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 669.186556] env[63418]: INFO nova.compute.manager [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: 938304f0-1b30-4848-a82d-eb5acf732677] Took 1.03 seconds to deallocate network for instance. [ 669.189711] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 669.559282] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.559282] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 669.561605] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.798s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.716767] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.066558] env[63418]: DEBUG nova.compute.utils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.071016] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 670.071402] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.110998] env[63418]: DEBUG nova.policy [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9a1603006084954b083a7f9d31c4918', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cd791635e444678ab02cb42f10a98ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 670.215672] env[63418]: INFO nova.scheduler.client.report [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Deleted allocations for instance 938304f0-1b30-4848-a82d-eb5acf732677 [ 670.430353] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Successfully created port: 9fd4ef6e-49be-4f7b-8236-5e563636b493 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.498559] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927fea29-b2e5-4e71-ba04-93617954c2b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.507930] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839a5595-7e71-4f03-8990-b9a70af7e4e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.542932] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75daae3a-6264-4b08-aa47-0737f851489d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.550166] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbf2ed6-8a03-42c1-b238-40b314fb9776 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.563145] env[63418]: DEBUG nova.compute.provider_tree [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.572259] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 670.729638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-88adedf7-3546-48bd-83b2-015d91be889f tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "938304f0-1b30-4848-a82d-eb5acf732677" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.779s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.068604] env[63418]: DEBUG nova.scheduler.client.report [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 671.120242] env[63418]: DEBUG nova.compute.manager [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Received event network-changed-9fd4ef6e-49be-4f7b-8236-5e563636b493 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 671.120718] env[63418]: DEBUG nova.compute.manager [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Refreshing instance network info cache due to event network-changed-9fd4ef6e-49be-4f7b-8236-5e563636b493. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 671.120718] env[63418]: DEBUG oslo_concurrency.lockutils [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] Acquiring lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.120875] env[63418]: DEBUG oslo_concurrency.lockutils [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] Acquired lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.121068] env[63418]: DEBUG nova.network.neutron [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Refreshing network info cache for port 9fd4ef6e-49be-4f7b-8236-5e563636b493 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 671.231333] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 671.301832] env[63418]: ERROR nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 671.301832] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.301832] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.301832] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.301832] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.301832] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.301832] env[63418]: ERROR nova.compute.manager raise self.value [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.301832] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.301832] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.301832] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.302376] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.302376] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.302376] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 671.302376] env[63418]: ERROR nova.compute.manager [ 671.302376] env[63418]: Traceback (most recent call last): [ 671.302376] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.302376] env[63418]: listener.cb(fileno) [ 671.302376] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.302376] env[63418]: result = function(*args, **kwargs) [ 671.302376] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.302376] env[63418]: return func(*args, **kwargs) [ 671.302376] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 671.302376] env[63418]: raise e [ 671.302376] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.302376] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 671.302376] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.302376] env[63418]: created_port_ids = self._update_ports_for_instance( [ 671.302376] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.302376] env[63418]: with excutils.save_and_reraise_exception(): [ 671.302376] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.302376] env[63418]: self.force_reraise() [ 671.302376] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.302376] env[63418]: raise self.value [ 671.302376] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.302376] env[63418]: updated_port = self._update_port( [ 671.302376] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.302376] env[63418]: _ensure_no_port_binding_failure(port) [ 671.302376] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.302376] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.303156] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 671.303156] env[63418]: Removing descriptor: 15 [ 671.573310] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.574034] env[63418]: ERROR nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Traceback (most recent call last): [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self.driver.spawn(context, instance, image_meta, [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] vm_ref = self.build_virtual_machine(instance, [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.574034] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] for vif in network_info: [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return self._sync_wrapper(fn, *args, **kwargs) [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self.wait() [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self[:] = self._gt.wait() [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return self._exit_event.wait() [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] result = hub.switch() [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.574455] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return self.greenlet.switch() [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] result = function(*args, **kwargs) [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] return func(*args, **kwargs) [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] raise e [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] nwinfo = self.network_api.allocate_for_instance( [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] created_port_ids = self._update_ports_for_instance( [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] with excutils.save_and_reraise_exception(): [ 671.574826] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] self.force_reraise() [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] raise self.value [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] updated_port = self._update_port( [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] _ensure_no_port_binding_failure(port) [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] raise exception.PortBindingFailed(port_id=port['id']) [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] nova.exception.PortBindingFailed: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. [ 671.575207] env[63418]: ERROR nova.compute.manager [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] [ 671.575684] env[63418]: DEBUG nova.compute.utils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.576052] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.571s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.578842] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Build of instance 76e02181-1ccd-43f9-9aaa-cf1918691705 was re-scheduled: Binding failed for port 40bca8e6-54c6-45ea-b1b3-f42428262085, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 671.580246] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 671.580246] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquiring lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.580246] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Acquired lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.580246] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.581996] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 671.610624] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.610624] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.612018] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.612018] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.612018] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.612018] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.612018] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.612463] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.612790] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.613102] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.613461] env[63418]: DEBUG nova.virt.hardware [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.614670] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c581389-70ca-4652-a50c-0270ad50c570 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.625573] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd08e08-d649-4934-af0c-bf3715426864 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.640829] env[63418]: ERROR nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Traceback (most recent call last): [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] yield resources [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self.driver.spawn(context, instance, image_meta, [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] vm_ref = self.build_virtual_machine(instance, [ 671.640829] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] for vif in network_info: [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] return self._sync_wrapper(fn, *args, **kwargs) [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self.wait() [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self[:] = self._gt.wait() [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] return self._exit_event.wait() [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 671.641240] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] current.throw(*self._exc) [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] result = function(*args, **kwargs) [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] return func(*args, **kwargs) [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] raise e [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] nwinfo = self.network_api.allocate_for_instance( [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] created_port_ids = self._update_ports_for_instance( [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] with excutils.save_and_reraise_exception(): [ 671.641602] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self.force_reraise() [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] raise self.value [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] updated_port = self._update_port( [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] _ensure_no_port_binding_failure(port) [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] raise exception.PortBindingFailed(port_id=port['id']) [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 671.641924] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] [ 671.641924] env[63418]: INFO nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Terminating instance [ 671.642524] env[63418]: DEBUG nova.network.neutron [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.722609] env[63418]: DEBUG nova.network.neutron [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.751803] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.100748] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.145126] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.146494] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquiring lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.224618] env[63418]: DEBUG oslo_concurrency.lockutils [req-6d705d6d-03df-4e44-ae56-3ccb41e9df78 req-83b84a62-e3a4-485f-bcf9-b7d82207481e service nova] Releasing lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.225221] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquired lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.225513] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.425044] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5343369e-1bf7-4ce5-a21a-5afcd65f07b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.431816] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93417a0-0bde-414b-b3e2-82306cb58430 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.461626] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa39bfb-18fc-4291-92d5-095cae6dc4ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.468835] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0048ffd1-d218-4598-95e5-944cefdd14a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.482020] env[63418]: DEBUG nova.compute.provider_tree [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.653045] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Releasing lock "refresh_cache-76e02181-1ccd-43f9-9aaa-cf1918691705" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.653045] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 672.653045] env[63418]: DEBUG nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 672.653286] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.669063] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.742364] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.811568] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.984770] env[63418]: DEBUG nova.scheduler.client.report [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 673.147451] env[63418]: DEBUG nova.compute.manager [req-418014bd-2844-4b37-af5f-42e230e3a1a2 req-f5a85e6e-0de8-494c-807b-4120d508125a service nova] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Received event network-vif-deleted-9fd4ef6e-49be-4f7b-8236-5e563636b493 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 673.172056] env[63418]: DEBUG nova.network.neutron [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.314028] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Releasing lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.314484] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 673.314704] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.315057] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-edccf9c0-c961-4c8c-8d51-67c7730de3b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.325079] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26240bb8-07cd-4a81-8517-d5cbe206c14b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.348275] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c590bbfe-1ab8-4e7a-a3aa-89e1f579d756 could not be found. [ 673.348528] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 673.348753] env[63418]: INFO nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Took 0.03 seconds to destroy the instance on the hypervisor. [ 673.348950] env[63418]: DEBUG oslo.service.loopingcall [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.349174] env[63418]: DEBUG nova.compute.manager [-] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 673.349266] env[63418]: DEBUG nova.network.neutron [-] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.364854] env[63418]: DEBUG nova.network.neutron [-] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.489484] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.490081] env[63418]: ERROR nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Traceback (most recent call last): [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self.driver.spawn(context, instance, image_meta, [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] vm_ref = self.build_virtual_machine(instance, [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.490081] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] for vif in network_info: [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] return self._sync_wrapper(fn, *args, **kwargs) [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self.wait() [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self[:] = self._gt.wait() [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] return self._exit_event.wait() [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] current.throw(*self._exc) [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.490443] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] result = function(*args, **kwargs) [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] return func(*args, **kwargs) [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] raise e [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] nwinfo = self.network_api.allocate_for_instance( [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] created_port_ids = self._update_ports_for_instance( [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] with excutils.save_and_reraise_exception(): [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] self.force_reraise() [ 673.490821] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] raise self.value [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] updated_port = self._update_port( [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] _ensure_no_port_binding_failure(port) [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] raise exception.PortBindingFailed(port_id=port['id']) [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] nova.exception.PortBindingFailed: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. [ 673.491208] env[63418]: ERROR nova.compute.manager [instance: b56238f3-3507-49e0-8630-5f5a093c1101] [ 673.491208] env[63418]: DEBUG nova.compute.utils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.492111] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.035s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.494913] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Build of instance b56238f3-3507-49e0-8630-5f5a093c1101 was re-scheduled: Binding failed for port 961af346-5a23-4010-9158-e5f874e5240a, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 673.495324] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 673.495544] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquiring lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.495705] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Acquired lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.495871] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.674508] env[63418]: INFO nova.compute.manager [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] [instance: 76e02181-1ccd-43f9-9aaa-cf1918691705] Took 1.02 seconds to deallocate network for instance. [ 673.868055] env[63418]: DEBUG nova.network.neutron [-] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.016141] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.092159] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.371362] env[63418]: INFO nova.compute.manager [-] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Took 1.02 seconds to deallocate network for instance. [ 674.374550] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742b7995-1154-491f-96e6-7c25befcccb5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.377371] env[63418]: DEBUG nova.compute.claims [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.377546] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.382594] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013b7c53-e28f-4280-b190-95ff9a069cba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.411547] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288f8ce7-95b7-4090-8bae-0b8e85074941 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.418463] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d26ef9-3171-45ba-953f-e5304eac1066 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.430984] env[63418]: DEBUG nova.compute.provider_tree [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.597260] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Releasing lock "refresh_cache-b56238f3-3507-49e0-8630-5f5a093c1101" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.597581] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 674.597624] env[63418]: DEBUG nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 674.597785] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.613193] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.703310] env[63418]: INFO nova.scheduler.client.report [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Deleted allocations for instance 76e02181-1ccd-43f9-9aaa-cf1918691705 [ 674.934461] env[63418]: DEBUG nova.scheduler.client.report [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 675.116042] env[63418]: DEBUG nova.network.neutron [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.210831] env[63418]: DEBUG oslo_concurrency.lockutils [None req-664acfcc-c991-4cc1-8774-802a39a31166 tempest-ListImageFiltersTestJSON-1818377368 tempest-ListImageFiltersTestJSON-1818377368-project-member] Lock "76e02181-1ccd-43f9-9aaa-cf1918691705" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.863s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.439669] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.947s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.440361] env[63418]: ERROR nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Traceback (most recent call last): [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self.driver.spawn(context, instance, image_meta, [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] vm_ref = self.build_virtual_machine(instance, [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.440361] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] for vif in network_info: [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] return self._sync_wrapper(fn, *args, **kwargs) [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self.wait() [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self[:] = self._gt.wait() [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] return self._exit_event.wait() [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] current.throw(*self._exc) [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.440729] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] result = function(*args, **kwargs) [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] return func(*args, **kwargs) [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] raise e [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] nwinfo = self.network_api.allocate_for_instance( [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] created_port_ids = self._update_ports_for_instance( [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] with excutils.save_and_reraise_exception(): [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] self.force_reraise() [ 675.441120] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] raise self.value [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] updated_port = self._update_port( [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] _ensure_no_port_binding_failure(port) [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] raise exception.PortBindingFailed(port_id=port['id']) [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] nova.exception.PortBindingFailed: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. [ 675.441559] env[63418]: ERROR nova.compute.manager [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] [ 675.441559] env[63418]: DEBUG nova.compute.utils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.442435] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.830s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.445390] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Build of instance 59442c77-234c-48c2-872d-85ed9bd8cb80 was re-scheduled: Binding failed for port cb30e2b7-0ce4-43cc-ac32-5782923ebd6e, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 675.445887] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 675.446190] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquiring lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.447029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Acquired lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.447029] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.620281] env[63418]: INFO nova.compute.manager [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] [instance: b56238f3-3507-49e0-8630-5f5a093c1101] Took 1.02 seconds to deallocate network for instance. [ 675.714097] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 675.970693] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.097563] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.237447] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.355630] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24151199-0bef-41ed-8031-f0435ce66674 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.363317] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4dd688-cf25-4d13-a5ab-b0b38a55e8d3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.393275] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b989fc0d-6782-4600-acb4-edc8c58f948e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.400818] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cc0e82-077a-43e3-9898-4906aaee6d82 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.415088] env[63418]: DEBUG nova.compute.provider_tree [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.602530] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Releasing lock "refresh_cache-59442c77-234c-48c2-872d-85ed9bd8cb80" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.602783] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 676.602964] env[63418]: DEBUG nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 676.603143] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.619180] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.655522] env[63418]: INFO nova.scheduler.client.report [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Deleted allocations for instance b56238f3-3507-49e0-8630-5f5a093c1101 [ 676.918778] env[63418]: DEBUG nova.scheduler.client.report [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 677.121920] env[63418]: DEBUG nova.network.neutron [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.163604] env[63418]: DEBUG oslo_concurrency.lockutils [None req-58b1afab-d6e1-4b3a-a575-38244e33819a tempest-ServersAdminTestJSON-1421506315 tempest-ServersAdminTestJSON-1421506315-project-member] Lock "b56238f3-3507-49e0-8630-5f5a093c1101" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.120s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.424285] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.424800] env[63418]: ERROR nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] Traceback (most recent call last): [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self.driver.spawn(context, instance, image_meta, [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] vm_ref = self.build_virtual_machine(instance, [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.424800] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] for vif in network_info: [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] return self._sync_wrapper(fn, *args, **kwargs) [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self.wait() [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self[:] = self._gt.wait() [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] return self._exit_event.wait() [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] current.throw(*self._exc) [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.425133] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] result = function(*args, **kwargs) [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] return func(*args, **kwargs) [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] raise e [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] nwinfo = self.network_api.allocate_for_instance( [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] created_port_ids = self._update_ports_for_instance( [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] with excutils.save_and_reraise_exception(): [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] self.force_reraise() [ 677.425553] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] raise self.value [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] updated_port = self._update_port( [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] _ensure_no_port_binding_failure(port) [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] raise exception.PortBindingFailed(port_id=port['id']) [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] nova.exception.PortBindingFailed: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. [ 677.425870] env[63418]: ERROR nova.compute.manager [instance: b24d095a-6814-413b-bf89-511e22df740f] [ 677.425870] env[63418]: DEBUG nova.compute.utils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.426856] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.799s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.428327] env[63418]: INFO nova.compute.claims [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.430940] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Build of instance b24d095a-6814-413b-bf89-511e22df740f was re-scheduled: Binding failed for port ade83678-a78b-4563-b665-2ec39a8ab330, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 677.431377] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 677.431621] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquiring lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.431784] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Acquired lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.431948] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.624183] env[63418]: INFO nova.compute.manager [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] [instance: 59442c77-234c-48c2-872d-85ed9bd8cb80] Took 1.02 seconds to deallocate network for instance. [ 677.666141] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 678.070902] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.190527] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.260353] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.660539] env[63418]: INFO nova.scheduler.client.report [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Deleted allocations for instance 59442c77-234c-48c2-872d-85ed9bd8cb80 [ 678.764796] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Releasing lock "refresh_cache-b24d095a-6814-413b-bf89-511e22df740f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.765090] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 678.765277] env[63418]: DEBUG nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 678.765449] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.785833] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.952524] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec28370d-62dc-4650-bf5f-16bcf3780a3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.960922] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7517ed-7129-4a37-8bc1-efc822c6e7b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.993323] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b87164-bc66-4f78-afe9-9a0d112f9021 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.001934] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7eca26-2107-42f1-beef-7fa1779c0dde {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.018603] env[63418]: DEBUG nova.compute.provider_tree [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.171032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3aeff048-95d9-4af8-bbc4-c0f2b2c843ae tempest-AttachInterfacesV270Test-1343056670 tempest-AttachInterfacesV270Test-1343056670-project-member] Lock "59442c77-234c-48c2-872d-85ed9bd8cb80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.611s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.291442] env[63418]: DEBUG nova.network.neutron [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.525371] env[63418]: DEBUG nova.scheduler.client.report [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 679.676793] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 679.794584] env[63418]: INFO nova.compute.manager [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] [instance: b24d095a-6814-413b-bf89-511e22df740f] Took 1.03 seconds to deallocate network for instance. [ 680.033699] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.603s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.033699] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 680.034088] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.483s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.205956] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.539402] env[63418]: DEBUG nova.compute.utils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.544381] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 680.544572] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.603517] env[63418]: DEBUG nova.policy [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63dab9a3fa4c407fb8b65602bc70c780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d3db19ea9a742f9a3144ee524078828', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.832232] env[63418]: INFO nova.scheduler.client.report [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Deleted allocations for instance b24d095a-6814-413b-bf89-511e22df740f [ 681.045586] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 681.069157] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c38c56-76f4-4bac-80b7-3f8cf28fc66b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.073685] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Successfully created port: 07a1d165-1209-42f0-b086-037ed171d4e4 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.081096] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d13102e-a34f-44d2-9a9e-fac0f88e672f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.115111] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b7e4cf-bceb-4feb-bcf3-75c73c306e92 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.124249] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5483cd-0b02-4a77-b28d-25343f5b3729 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.140369] env[63418]: DEBUG nova.compute.provider_tree [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.346471] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83567000-ac06-4267-b134-f10993656302 tempest-FloatingIPsAssociationTestJSON-1189590740 tempest-FloatingIPsAssociationTestJSON-1189590740-project-member] Lock "b24d095a-6814-413b-bf89-511e22df740f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.143s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.643444] env[63418]: DEBUG nova.scheduler.client.report [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 681.850207] env[63418]: DEBUG nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 681.988155] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "27290558-6c58-414d-bd53-c4a686890721" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.988828] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "27290558-6c58-414d-bd53-c4a686890721" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.060973] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 682.114832] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.114832] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.114832] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.115076] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.115076] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.115076] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.115076] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.115191] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.115352] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.115441] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.115605] env[63418]: DEBUG nova.virt.hardware [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.116636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a2e397-a884-472b-910c-47d252683fd1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.125326] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c812806-e715-490f-9b72-ffd97104b7e7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.148849] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.116s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.149501] env[63418]: ERROR nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Traceback (most recent call last): [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self.driver.spawn(context, instance, image_meta, [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] vm_ref = self.build_virtual_machine(instance, [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.149501] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] for vif in network_info: [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] return self._sync_wrapper(fn, *args, **kwargs) [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self.wait() [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self[:] = self._gt.wait() [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] return self._exit_event.wait() [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] current.throw(*self._exc) [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.149879] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] result = function(*args, **kwargs) [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] return func(*args, **kwargs) [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] raise e [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] nwinfo = self.network_api.allocate_for_instance( [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] created_port_ids = self._update_ports_for_instance( [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] with excutils.save_and_reraise_exception(): [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] self.force_reraise() [ 682.150214] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] raise self.value [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] updated_port = self._update_port( [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] _ensure_no_port_binding_failure(port) [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] raise exception.PortBindingFailed(port_id=port['id']) [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] nova.exception.PortBindingFailed: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. [ 682.150524] env[63418]: ERROR nova.compute.manager [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] [ 682.150524] env[63418]: DEBUG nova.compute.utils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.151953] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Build of instance 601f72f4-05bc-4d73-9b57-612cad2d8c62 was re-scheduled: Binding failed for port d325e415-e7a8-482f-aa6c-4eb998861d38, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 682.152421] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 682.152603] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquiring lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.152751] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Acquired lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.152905] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.153940] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.395s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.165571] env[63418]: ERROR nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 682.165571] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 682.165571] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.165571] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.165571] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.165571] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.165571] env[63418]: ERROR nova.compute.manager raise self.value [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.165571] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 682.165571] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.165571] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 682.166453] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.166453] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 682.166453] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 682.166453] env[63418]: ERROR nova.compute.manager [ 682.166453] env[63418]: Traceback (most recent call last): [ 682.166453] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 682.166453] env[63418]: listener.cb(fileno) [ 682.166453] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.166453] env[63418]: result = function(*args, **kwargs) [ 682.166453] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.166453] env[63418]: return func(*args, **kwargs) [ 682.166453] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 682.166453] env[63418]: raise e [ 682.166453] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 682.166453] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 682.166453] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.166453] env[63418]: created_port_ids = self._update_ports_for_instance( [ 682.166453] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.166453] env[63418]: with excutils.save_and_reraise_exception(): [ 682.166453] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.166453] env[63418]: self.force_reraise() [ 682.166453] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.166453] env[63418]: raise self.value [ 682.166453] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.166453] env[63418]: updated_port = self._update_port( [ 682.166453] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.166453] env[63418]: _ensure_no_port_binding_failure(port) [ 682.166453] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.166453] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 682.167200] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 682.167200] env[63418]: Removing descriptor: 15 [ 682.167200] env[63418]: ERROR nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Traceback (most recent call last): [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] yield resources [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self.driver.spawn(context, instance, image_meta, [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.167200] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] vm_ref = self.build_virtual_machine(instance, [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] for vif in network_info: [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return self._sync_wrapper(fn, *args, **kwargs) [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self.wait() [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self[:] = self._gt.wait() [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return self._exit_event.wait() [ 682.167516] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] result = hub.switch() [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return self.greenlet.switch() [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] result = function(*args, **kwargs) [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return func(*args, **kwargs) [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] raise e [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] nwinfo = self.network_api.allocate_for_instance( [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.168588] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] created_port_ids = self._update_ports_for_instance( [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] with excutils.save_and_reraise_exception(): [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self.force_reraise() [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] raise self.value [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] updated_port = self._update_port( [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] _ensure_no_port_binding_failure(port) [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.169064] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] raise exception.PortBindingFailed(port_id=port['id']) [ 682.169918] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 682.169918] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] [ 682.169918] env[63418]: INFO nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Terminating instance [ 682.181940] env[63418]: DEBUG nova.compute.manager [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Received event network-changed-07a1d165-1209-42f0-b086-037ed171d4e4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 682.181940] env[63418]: DEBUG nova.compute.manager [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Refreshing instance network info cache due to event network-changed-07a1d165-1209-42f0-b086-037ed171d4e4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 682.181940] env[63418]: DEBUG oslo_concurrency.lockutils [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] Acquiring lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.181940] env[63418]: DEBUG oslo_concurrency.lockutils [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] Acquired lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.181940] env[63418]: DEBUG nova.network.neutron [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Refreshing network info cache for port 07a1d165-1209-42f0-b086-037ed171d4e4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 682.377579] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.674967] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquiring lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.685172] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.722965] env[63418]: DEBUG nova.network.neutron [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.815539] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.876452] env[63418]: DEBUG nova.network.neutron [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.098836] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafd832b-d8ee-46b2-bb73-8492f81278ef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.106310] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7142f954-9dbb-4ac9-87ab-e79b0e98ea60 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.136190] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b33d1c-505f-4e95-8246-095a39c95aae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.144160] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ac6fb6-2a0f-4f76-89d5-ce238455e325 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.157792] env[63418]: DEBUG nova.compute.provider_tree [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.318349] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Releasing lock "refresh_cache-601f72f4-05bc-4d73-9b57-612cad2d8c62" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.318601] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 683.318785] env[63418]: DEBUG nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 683.319106] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.336675] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.381823] env[63418]: DEBUG oslo_concurrency.lockutils [req-2dbdaf5a-921f-48c8-a6c8-016b0aca7c8b req-8076ba01-8cc5-426b-aba3-cbf70cad3856 service nova] Releasing lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.382280] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquired lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.382485] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.661261] env[63418]: DEBUG nova.scheduler.client.report [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 683.840136] env[63418]: DEBUG nova.network.neutron [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.907956] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.027212] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.167553] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.168228] env[63418]: ERROR nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Traceback (most recent call last): [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self.driver.spawn(context, instance, image_meta, [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] vm_ref = self.build_virtual_machine(instance, [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.168228] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] for vif in network_info: [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] return self._sync_wrapper(fn, *args, **kwargs) [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self.wait() [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self[:] = self._gt.wait() [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] return self._exit_event.wait() [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] current.throw(*self._exc) [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.168513] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] result = function(*args, **kwargs) [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] return func(*args, **kwargs) [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] raise e [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] nwinfo = self.network_api.allocate_for_instance( [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] created_port_ids = self._update_ports_for_instance( [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] with excutils.save_and_reraise_exception(): [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] self.force_reraise() [ 684.168844] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] raise self.value [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] updated_port = self._update_port( [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] _ensure_no_port_binding_failure(port) [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] raise exception.PortBindingFailed(port_id=port['id']) [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] nova.exception.PortBindingFailed: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. [ 684.169288] env[63418]: ERROR nova.compute.manager [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] [ 684.169288] env[63418]: DEBUG nova.compute.utils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.170471] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.454s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.174290] env[63418]: INFO nova.compute.claims [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.174812] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Build of instance 6d47c52c-c1fb-4ade-a3ba-dc393716eb79 was re-scheduled: Binding failed for port cc684a4f-1b55-4017-8a0f-5787cd0654bc, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 684.175261] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 684.175480] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.175633] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.175789] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.254270] env[63418]: DEBUG nova.compute.manager [req-4a0effd9-b51b-4117-b98d-c513e718b7a7 req-26d2830f-fecd-4981-a918-d164a6bcad23 service nova] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Received event network-vif-deleted-07a1d165-1209-42f0-b086-037ed171d4e4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 684.344996] env[63418]: INFO nova.compute.manager [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] [instance: 601f72f4-05bc-4d73-9b57-612cad2d8c62] Took 1.03 seconds to deallocate network for instance. [ 684.530128] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Releasing lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.530128] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 684.530128] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.530289] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb406fab-f172-41c7-abfe-70bf65d58bf8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.539822] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4cfbab-c703-4f9a-99d1-f0a4661dcf5d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.561863] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6d62d1b-0333-40ac-92e8-a798e58d495c could not be found. [ 684.562099] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.562287] env[63418]: INFO nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 684.562525] env[63418]: DEBUG oslo.service.loopingcall [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.562756] env[63418]: DEBUG nova.compute.manager [-] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 684.562871] env[63418]: DEBUG nova.network.neutron [-] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.587890] env[63418]: DEBUG nova.network.neutron [-] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.701633] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.816009] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.090195] env[63418]: DEBUG nova.network.neutron [-] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.318237] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-6d47c52c-c1fb-4ade-a3ba-dc393716eb79" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.318502] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 685.318699] env[63418]: DEBUG nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 685.318863] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.342986] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.382494] env[63418]: INFO nova.scheduler.client.report [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Deleted allocations for instance 601f72f4-05bc-4d73-9b57-612cad2d8c62 [ 685.596485] env[63418]: INFO nova.compute.manager [-] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Took 1.03 seconds to deallocate network for instance. [ 685.598754] env[63418]: DEBUG nova.compute.claims [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 685.598993] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.675503] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405f84cc-8ee6-4608-ba0c-41b1209ea445 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.683515] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3ab05f-12bc-47a3-ab17-4b849a73de48 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.716855] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edde9ee-c89d-4abf-86a1-23578787d2eb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.724632] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd0741d-1686-419b-961f-25700120eb0d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.737704] env[63418]: DEBUG nova.compute.provider_tree [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.849110] env[63418]: DEBUG nova.network.neutron [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.893493] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52bcc151-405c-4956-8488-3f182bfb3e82 tempest-ServersTestBootFromVolume-1301034763 tempest-ServersTestBootFromVolume-1301034763-project-member] Lock "601f72f4-05bc-4d73-9b57-612cad2d8c62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.640s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.243177] env[63418]: DEBUG nova.scheduler.client.report [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 686.351896] env[63418]: INFO nova.compute.manager [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 6d47c52c-c1fb-4ade-a3ba-dc393716eb79] Took 1.03 seconds to deallocate network for instance. [ 686.399044] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 686.747598] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.748178] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 686.751927] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.753320] env[63418]: INFO nova.compute.claims [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.922586] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.255647] env[63418]: DEBUG nova.compute.utils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.255647] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 687.256532] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 687.337502] env[63418]: DEBUG nova.policy [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e79308c25de04876964c184c5af674ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a389512a583d499099e898eb0f560016', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.402025] env[63418]: INFO nova.scheduler.client.report [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted allocations for instance 6d47c52c-c1fb-4ade-a3ba-dc393716eb79 [ 687.709849] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Successfully created port: 75df9a9b-8bf5-414c-99c5-f3897ca7e448 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.759874] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 687.914913] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dbab6af0-4a81-494b-8bd3-f4b5defd157c tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "6d47c52c-c1fb-4ade-a3ba-dc393716eb79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.242s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.200012] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527ee892-160b-4cd9-ba5b-a98e8839b06b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.208832] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a99fe4-b2e2-4baa-95b1-09121b56b5ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.239707] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40745cd-f4c7-4e96-ae50-42a18a81ba1f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.247123] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b8df56-6a4d-46d5-91f6-aab7c7e9edb8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.262291] env[63418]: DEBUG nova.compute.provider_tree [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.419697] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 688.771923] env[63418]: DEBUG nova.scheduler.client.report [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 688.779022] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 688.810456] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.813326] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.813872] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.813872] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.813872] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.814013] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.814599] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.814945] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.815216] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.815446] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.815704] env[63418]: DEBUG nova.virt.hardware [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.817287] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba19087-87fb-45d3-95f7-bb187a7c7743 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.829101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274a54b5-5e58-4e48-a653-0cbb0ecf3263 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.947469] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.095928] env[63418]: ERROR nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 689.095928] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.095928] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.095928] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.095928] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.095928] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.095928] env[63418]: ERROR nova.compute.manager raise self.value [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.095928] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.095928] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.095928] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.096421] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.096421] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.096421] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 689.096421] env[63418]: ERROR nova.compute.manager [ 689.096421] env[63418]: Traceback (most recent call last): [ 689.096421] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.096421] env[63418]: listener.cb(fileno) [ 689.096421] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.096421] env[63418]: result = function(*args, **kwargs) [ 689.096421] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.096421] env[63418]: return func(*args, **kwargs) [ 689.096421] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 689.096421] env[63418]: raise e [ 689.096421] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.096421] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 689.096421] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.096421] env[63418]: created_port_ids = self._update_ports_for_instance( [ 689.096421] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.096421] env[63418]: with excutils.save_and_reraise_exception(): [ 689.096421] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.096421] env[63418]: self.force_reraise() [ 689.096421] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.096421] env[63418]: raise self.value [ 689.096421] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.096421] env[63418]: updated_port = self._update_port( [ 689.096421] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.096421] env[63418]: _ensure_no_port_binding_failure(port) [ 689.096421] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.096421] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.097529] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 689.097529] env[63418]: Removing descriptor: 16 [ 689.097529] env[63418]: ERROR nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Traceback (most recent call last): [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] yield resources [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self.driver.spawn(context, instance, image_meta, [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.097529] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] vm_ref = self.build_virtual_machine(instance, [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] for vif in network_info: [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return self._sync_wrapper(fn, *args, **kwargs) [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self.wait() [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self[:] = self._gt.wait() [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return self._exit_event.wait() [ 689.098254] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] result = hub.switch() [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return self.greenlet.switch() [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] result = function(*args, **kwargs) [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return func(*args, **kwargs) [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] raise e [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] nwinfo = self.network_api.allocate_for_instance( [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.098633] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] created_port_ids = self._update_ports_for_instance( [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] with excutils.save_and_reraise_exception(): [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self.force_reraise() [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] raise self.value [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] updated_port = self._update_port( [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] _ensure_no_port_binding_failure(port) [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.099346] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] raise exception.PortBindingFailed(port_id=port['id']) [ 689.100026] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 689.100026] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] [ 689.100026] env[63418]: INFO nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Terminating instance [ 689.217733] env[63418]: DEBUG nova.compute.manager [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Received event network-changed-75df9a9b-8bf5-414c-99c5-f3897ca7e448 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 689.217845] env[63418]: DEBUG nova.compute.manager [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Refreshing instance network info cache due to event network-changed-75df9a9b-8bf5-414c-99c5-f3897ca7e448. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 689.218059] env[63418]: DEBUG oslo_concurrency.lockutils [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] Acquiring lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.218205] env[63418]: DEBUG oslo_concurrency.lockutils [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] Acquired lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.218705] env[63418]: DEBUG nova.network.neutron [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Refreshing network info cache for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 689.282085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.282654] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 689.285946] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.908s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.600551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquiring lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.700405] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.700660] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.740990] env[63418]: DEBUG nova.network.neutron [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.790624] env[63418]: DEBUG nova.compute.utils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.798302] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 689.798302] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 689.840500] env[63418]: DEBUG nova.network.neutron [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.862091] env[63418]: DEBUG nova.policy [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc5c0fadf0b54b71957a87d56ea10acc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c828658cd364400a9207d0c90f3ae1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.145193] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Successfully created port: 1295a3e6-1a03-49d8-bfaa-fe456210135f {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.212408] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f120eb04-a101-4c06-90ee-8036555f2544 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.221622] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4b01e1-26c6-4721-adf7-d23687022993 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.252548] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34134c1b-8a71-481b-a222-ddd34341f715 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.259906] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a68bc25-80f0-4064-91d7-c7880995914f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.273113] env[63418]: DEBUG nova.compute.provider_tree [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.296126] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 690.344640] env[63418]: DEBUG oslo_concurrency.lockutils [req-54cb3253-d445-4645-841a-36b849b32f13 req-57640652-e4f3-4dde-bf68-bd5e584a48a5 service nova] Releasing lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.345401] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquired lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.345401] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.780175] env[63418]: DEBUG nova.scheduler.client.report [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 690.872894] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.055853] env[63418]: DEBUG nova.compute.manager [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Received event network-changed-1295a3e6-1a03-49d8-bfaa-fe456210135f {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 691.056070] env[63418]: DEBUG nova.compute.manager [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Refreshing instance network info cache due to event network-changed-1295a3e6-1a03-49d8-bfaa-fe456210135f. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 691.056280] env[63418]: DEBUG oslo_concurrency.lockutils [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] Acquiring lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.056579] env[63418]: DEBUG oslo_concurrency.lockutils [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] Acquired lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.056579] env[63418]: DEBUG nova.network.neutron [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Refreshing network info cache for port 1295a3e6-1a03-49d8-bfaa-fe456210135f {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 691.075565] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.286372] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.287108] env[63418]: ERROR nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Traceback (most recent call last): [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self.driver.spawn(context, instance, image_meta, [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] vm_ref = self.build_virtual_machine(instance, [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.287108] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] for vif in network_info: [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] return self._sync_wrapper(fn, *args, **kwargs) [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self.wait() [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self[:] = self._gt.wait() [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] return self._exit_event.wait() [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] current.throw(*self._exc) [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.287436] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] result = function(*args, **kwargs) [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] return func(*args, **kwargs) [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] raise e [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] nwinfo = self.network_api.allocate_for_instance( [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] created_port_ids = self._update_ports_for_instance( [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] with excutils.save_and_reraise_exception(): [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] self.force_reraise() [ 691.287928] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] raise self.value [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] updated_port = self._update_port( [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] _ensure_no_port_binding_failure(port) [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] raise exception.PortBindingFailed(port_id=port['id']) [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] nova.exception.PortBindingFailed: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. [ 691.288297] env[63418]: ERROR nova.compute.manager [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] [ 691.288297] env[63418]: DEBUG nova.compute.utils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.289696] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Build of instance c590bbfe-1ab8-4e7a-a3aa-89e1f579d756 was re-scheduled: Binding failed for port 9fd4ef6e-49be-4f7b-8236-5e563636b493, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 691.290365] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 691.290600] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquiring lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.290752] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Acquired lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.290908] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.292228] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.055s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.297020] env[63418]: INFO nova.compute.claims [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.304877] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 691.335099] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.335375] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.335564] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.335731] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.335893] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.336654] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.336958] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.337252] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.337357] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.338361] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.338361] env[63418]: DEBUG nova.virt.hardware [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.338676] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af475db7-cdaa-4f34-9050-241990226fb5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.350436] env[63418]: ERROR nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 691.350436] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.350436] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.350436] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.350436] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.350436] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.350436] env[63418]: ERROR nova.compute.manager raise self.value [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.350436] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.350436] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.350436] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.350968] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.350968] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.350968] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 691.350968] env[63418]: ERROR nova.compute.manager [ 691.350968] env[63418]: Traceback (most recent call last): [ 691.350968] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.350968] env[63418]: listener.cb(fileno) [ 691.350968] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.350968] env[63418]: result = function(*args, **kwargs) [ 691.350968] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.350968] env[63418]: return func(*args, **kwargs) [ 691.350968] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.350968] env[63418]: raise e [ 691.350968] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.350968] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 691.350968] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.350968] env[63418]: created_port_ids = self._update_ports_for_instance( [ 691.350968] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.350968] env[63418]: with excutils.save_and_reraise_exception(): [ 691.350968] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.350968] env[63418]: self.force_reraise() [ 691.350968] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.350968] env[63418]: raise self.value [ 691.350968] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.350968] env[63418]: updated_port = self._update_port( [ 691.350968] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.350968] env[63418]: _ensure_no_port_binding_failure(port) [ 691.350968] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.350968] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.351807] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 691.351807] env[63418]: Removing descriptor: 15 [ 691.351932] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb321b5c-6066-486e-8f52-ea211ee286c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.368339] env[63418]: ERROR nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Traceback (most recent call last): [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] yield resources [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self.driver.spawn(context, instance, image_meta, [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] vm_ref = self.build_virtual_machine(instance, [ 691.368339] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] for vif in network_info: [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] return self._sync_wrapper(fn, *args, **kwargs) [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self.wait() [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self[:] = self._gt.wait() [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] return self._exit_event.wait() [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.368759] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] current.throw(*self._exc) [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] result = function(*args, **kwargs) [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] return func(*args, **kwargs) [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] raise e [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] nwinfo = self.network_api.allocate_for_instance( [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] created_port_ids = self._update_ports_for_instance( [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] with excutils.save_and_reraise_exception(): [ 691.370204] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self.force_reraise() [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] raise self.value [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] updated_port = self._update_port( [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] _ensure_no_port_binding_failure(port) [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] raise exception.PortBindingFailed(port_id=port['id']) [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 691.370631] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] [ 691.370631] env[63418]: INFO nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Terminating instance [ 691.517196] env[63418]: DEBUG nova.compute.manager [req-233e03e7-8785-4298-bebc-44096574712f req-5b42d52b-1b77-4cd0-94f7-22f52e7edea1 service nova] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Received event network-vif-deleted-75df9a9b-8bf5-414c-99c5-f3897ca7e448 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 691.577681] env[63418]: DEBUG nova.network.neutron [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.581247] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Releasing lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.582088] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 691.582465] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.583075] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9380608-6b0d-41c9-8998-de65259d800f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.596674] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2ee562-ce42-4f06-ad5d-44c53e8a9ab9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.616180] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e39c8ef-103a-45f4-8630-bfd51b38445d could not be found. [ 691.616437] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.616617] env[63418]: INFO nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 691.616863] env[63418]: DEBUG oslo.service.loopingcall [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.617167] env[63418]: DEBUG nova.compute.manager [-] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 691.617207] env[63418]: DEBUG nova.network.neutron [-] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.644985] env[63418]: DEBUG nova.network.neutron [-] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.738005] env[63418]: DEBUG nova.network.neutron [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.824693] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.873112] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.968469] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.147992] env[63418]: DEBUG nova.network.neutron [-] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.241252] env[63418]: DEBUG oslo_concurrency.lockutils [req-6fe10b11-de77-499c-80e7-d5bd3436cd65 req-fbf37a77-4b35-4de9-9050-3835e93644db service nova] Releasing lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.241684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.241949] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.471583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Releasing lock "refresh_cache-c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.471583] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 692.472529] env[63418]: DEBUG nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 692.472529] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.493798] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.653255] env[63418]: INFO nova.compute.manager [-] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Took 1.03 seconds to deallocate network for instance. [ 692.654245] env[63418]: DEBUG nova.compute.claims [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.654621] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.711653] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2d500f-373c-4f4a-a14c-6510961e58fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.721281] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50458028-83c2-427d-8d06-805aff29264c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.756994] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889f3acf-ce33-4011-937d-7366e0c952d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.766067] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5db440-fe38-4c8c-b646-acd3e8d48091 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.780774] env[63418]: DEBUG nova.compute.provider_tree [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.782713] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.788697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.788910] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.899632] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.997327] env[63418]: DEBUG nova.network.neutron [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.201092] env[63418]: DEBUG nova.compute.manager [req-300541d0-74c0-4d67-aef5-60ca08c0557c req-a6d38671-1b2b-4f82-86a4-e83f0a3d17a6 service nova] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Received event network-vif-deleted-1295a3e6-1a03-49d8-bfaa-fe456210135f {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 693.288035] env[63418]: DEBUG nova.scheduler.client.report [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 693.402140] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Releasing lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.402506] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 693.402698] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.403007] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfb3bd9d-bc08-44a3-81e4-7f6838df8b62 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.413532] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b153e21-8b0a-4e81-a1c5-a4cb07157c63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.435171] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69651fab-08a7-4832-8c91-2fa01d186860 could not be found. [ 693.435415] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.435596] env[63418]: INFO nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Took 0.03 seconds to destroy the instance on the hypervisor. [ 693.435835] env[63418]: DEBUG oslo.service.loopingcall [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.436072] env[63418]: DEBUG nova.compute.manager [-] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 693.436164] env[63418]: DEBUG nova.network.neutron [-] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.453420] env[63418]: DEBUG nova.network.neutron [-] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.502788] env[63418]: INFO nova.compute.manager [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] [instance: c590bbfe-1ab8-4e7a-a3aa-89e1f579d756] Took 1.03 seconds to deallocate network for instance. [ 693.792532] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.793073] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 693.795705] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.605s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.797515] env[63418]: INFO nova.compute.claims [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.955108] env[63418]: DEBUG nova.network.neutron [-] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.301958] env[63418]: DEBUG nova.compute.utils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 694.305086] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 694.305208] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 694.352853] env[63418]: DEBUG nova.policy [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5736c3bfbb4806b6aaef5a6e05ca7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34143fc5e0144a8fb736d3deca6cc20b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.462561] env[63418]: INFO nova.compute.manager [-] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Took 1.03 seconds to deallocate network for instance. [ 694.465075] env[63418]: DEBUG nova.compute.claims [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.465253] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.469893] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.470126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.534438] env[63418]: INFO nova.scheduler.client.report [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Deleted allocations for instance c590bbfe-1ab8-4e7a-a3aa-89e1f579d756 [ 694.668596] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Successfully created port: c6429985-1dff-4d47-967a-a20c8738a942 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.805930] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 694.943585] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Successfully created port: 39caf1da-6760-4635-a7ca-c67467dfb5b0 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.046661] env[63418]: DEBUG oslo_concurrency.lockutils [None req-475ddd81-3ea5-48e6-a9b7-40647165c91b tempest-AttachInterfacesUnderV243Test-365559923 tempest-AttachInterfacesUnderV243Test-365559923-project-member] Lock "c590bbfe-1ab8-4e7a-a3aa-89e1f579d756" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.564s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.246544] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c665d3dd-f08f-4631-845f-f0021571d6e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.254866] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1e2c10-f5f0-4c3b-bbd4-57922f45a1f9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.289412] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e2b622-6a2f-4fcc-81b3-d0193a06726b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.298168] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb47e34-f14d-464c-9cbc-494a68bce94e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.322833] env[63418]: DEBUG nova.compute.provider_tree [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.549114] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 695.632015] env[63418]: DEBUG nova.compute.manager [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Received event network-changed-c6429985-1dff-4d47-967a-a20c8738a942 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 695.632228] env[63418]: DEBUG nova.compute.manager [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Refreshing instance network info cache due to event network-changed-c6429985-1dff-4d47-967a-a20c8738a942. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 695.632437] env[63418]: DEBUG oslo_concurrency.lockutils [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] Acquiring lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.632638] env[63418]: DEBUG oslo_concurrency.lockutils [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] Acquired lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.632719] env[63418]: DEBUG nova.network.neutron [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Refreshing network info cache for port c6429985-1dff-4d47-967a-a20c8738a942 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 695.807991] env[63418]: ERROR nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 695.807991] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.807991] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.807991] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.807991] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.807991] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.807991] env[63418]: ERROR nova.compute.manager raise self.value [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.807991] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.807991] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.807991] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.808534] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.808534] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.808534] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 695.808534] env[63418]: ERROR nova.compute.manager [ 695.808534] env[63418]: Traceback (most recent call last): [ 695.808534] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.808534] env[63418]: listener.cb(fileno) [ 695.808534] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.808534] env[63418]: result = function(*args, **kwargs) [ 695.808534] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.808534] env[63418]: return func(*args, **kwargs) [ 695.808534] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.808534] env[63418]: raise e [ 695.808534] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.808534] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 695.808534] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.808534] env[63418]: created_port_ids = self._update_ports_for_instance( [ 695.808534] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.808534] env[63418]: with excutils.save_and_reraise_exception(): [ 695.808534] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.808534] env[63418]: self.force_reraise() [ 695.808534] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.808534] env[63418]: raise self.value [ 695.808534] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.808534] env[63418]: updated_port = self._update_port( [ 695.808534] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.808534] env[63418]: _ensure_no_port_binding_failure(port) [ 695.808534] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.808534] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.809486] env[63418]: nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 695.809486] env[63418]: Removing descriptor: 15 [ 695.828349] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 695.831011] env[63418]: DEBUG nova.scheduler.client.report [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 695.862057] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.862312] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.862467] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.862646] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.862815] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.862975] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.863306] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.863525] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.863746] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.863940] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.864154] env[63418]: DEBUG nova.virt.hardware [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.865081] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f7fdc5-9154-4f4c-8c12-0a25ab513b2a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.874068] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038cc1f0-68c6-4497-88a6-70c1c5a5199e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.888946] env[63418]: ERROR nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Traceback (most recent call last): [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] yield resources [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self.driver.spawn(context, instance, image_meta, [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] vm_ref = self.build_virtual_machine(instance, [ 695.888946] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] for vif in network_info: [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] return self._sync_wrapper(fn, *args, **kwargs) [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self.wait() [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self[:] = self._gt.wait() [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] return self._exit_event.wait() [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.889348] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] current.throw(*self._exc) [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] result = function(*args, **kwargs) [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] return func(*args, **kwargs) [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] raise e [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] nwinfo = self.network_api.allocate_for_instance( [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] created_port_ids = self._update_ports_for_instance( [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] with excutils.save_and_reraise_exception(): [ 695.889702] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self.force_reraise() [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] raise self.value [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] updated_port = self._update_port( [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] _ensure_no_port_binding_failure(port) [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] raise exception.PortBindingFailed(port_id=port['id']) [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 695.890079] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] [ 695.890079] env[63418]: INFO nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Terminating instance [ 696.082062] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.169245] env[63418]: DEBUG nova.network.neutron [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.253269] env[63418]: DEBUG nova.network.neutron [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.337672] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.338296] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 696.342512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.137s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.344378] env[63418]: INFO nova.compute.claims [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.393562] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.756557] env[63418]: DEBUG oslo_concurrency.lockutils [req-e215a2fb-83d6-43bf-b8d4-65ade1456ed6 req-0cd310ac-88aa-4369-a5cf-1fdc1179d63a service nova] Releasing lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.756983] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquired lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.757477] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.849755] env[63418]: DEBUG nova.compute.utils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.851173] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 696.851376] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.895248] env[63418]: DEBUG nova.policy [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '124a501416ee4bb49394c2aefdf98e95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0525c553595147feaf2b3aa1c3e9b02f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.195415] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Successfully created port: db5be14b-80a3-4322-b1a0-e6354d1eb8e4 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.330618] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.358336] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 697.506048] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.705329] env[63418]: DEBUG nova.compute.manager [req-00691aae-bb2e-4717-a4f5-f575ab56683a req-5ff7f2c7-949d-426a-a52a-267c3f7a4704 service nova] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Received event network-vif-deleted-c6429985-1dff-4d47-967a-a20c8738a942 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 697.841035] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f953d376-f127-449b-86c1-19863beece4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.848903] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbab2d1-3044-4a3f-bbeb-283b5c4d474d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.887160] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a9c891-3f53-454d-849b-fbbfa6aec1fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.891809] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b44cea-6e23-4b22-8e23-80a3d037532c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.905439] env[63418]: DEBUG nova.compute.provider_tree [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.010007] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Releasing lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.010468] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 698.010665] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 698.011394] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24732987-40e8-4725-aacf-f9d31f1ccc07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.020532] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c955d2-d345-4208-9018-96518150fb56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.043459] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2edbeef4-be3b-4a1d-8895-8410b71b6b18 could not be found. [ 698.043548] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 698.043838] env[63418]: INFO nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Took 0.03 seconds to destroy the instance on the hypervisor. [ 698.044018] env[63418]: DEBUG oslo.service.loopingcall [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.044352] env[63418]: DEBUG nova.compute.manager [-] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 698.044413] env[63418]: DEBUG nova.network.neutron [-] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.084617] env[63418]: DEBUG nova.network.neutron [-] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.090889] env[63418]: ERROR nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 698.090889] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 698.090889] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.090889] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.090889] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.090889] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.090889] env[63418]: ERROR nova.compute.manager raise self.value [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.090889] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.090889] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.090889] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.092368] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.092368] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.092368] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 698.092368] env[63418]: ERROR nova.compute.manager [ 698.092368] env[63418]: Traceback (most recent call last): [ 698.092368] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.092368] env[63418]: listener.cb(fileno) [ 698.092368] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.092368] env[63418]: result = function(*args, **kwargs) [ 698.092368] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.092368] env[63418]: return func(*args, **kwargs) [ 698.092368] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 698.092368] env[63418]: raise e [ 698.092368] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 698.092368] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 698.092368] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.092368] env[63418]: created_port_ids = self._update_ports_for_instance( [ 698.092368] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.092368] env[63418]: with excutils.save_and_reraise_exception(): [ 698.092368] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.092368] env[63418]: self.force_reraise() [ 698.092368] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.092368] env[63418]: raise self.value [ 698.092368] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.092368] env[63418]: updated_port = self._update_port( [ 698.092368] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.092368] env[63418]: _ensure_no_port_binding_failure(port) [ 698.092368] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.092368] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.094220] env[63418]: nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 698.094220] env[63418]: Removing descriptor: 15 [ 698.388214] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 698.408703] env[63418]: DEBUG nova.scheduler.client.report [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 698.427026] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.427026] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.427026] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.427268] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.427522] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.427831] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.428179] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.429086] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.429434] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.429922] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.430258] env[63418]: DEBUG nova.virt.hardware [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.433234] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354aa628-e316-487e-ba07-1e37e320bd77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.442540] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d816d9e2-439f-4a8c-bfa9-fadcba429e27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.459324] env[63418]: ERROR nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Traceback (most recent call last): [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] yield resources [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self.driver.spawn(context, instance, image_meta, [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] vm_ref = self.build_virtual_machine(instance, [ 698.459324] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] for vif in network_info: [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] return self._sync_wrapper(fn, *args, **kwargs) [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self.wait() [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self[:] = self._gt.wait() [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] return self._exit_event.wait() [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 698.459954] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] current.throw(*self._exc) [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] result = function(*args, **kwargs) [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] return func(*args, **kwargs) [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] raise e [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] nwinfo = self.network_api.allocate_for_instance( [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] created_port_ids = self._update_ports_for_instance( [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] with excutils.save_and_reraise_exception(): [ 698.460460] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self.force_reraise() [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] raise self.value [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] updated_port = self._update_port( [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] _ensure_no_port_binding_failure(port) [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] raise exception.PortBindingFailed(port_id=port['id']) [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 698.460913] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] [ 698.460913] env[63418]: INFO nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Terminating instance [ 698.794825] env[63418]: DEBUG nova.network.neutron [-] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.916609] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.917234] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 698.920023] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.543s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.923248] env[63418]: INFO nova.compute.claims [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.964161] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquiring lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.964355] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquired lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.964565] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.297095] env[63418]: INFO nova.compute.manager [-] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Took 1.25 seconds to deallocate network for instance. [ 699.300199] env[63418]: DEBUG nova.compute.claims [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 699.300291] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.427919] env[63418]: DEBUG nova.compute.utils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.434790] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 699.434790] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.484129] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.538237] env[63418]: DEBUG nova.policy [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e14c925f44747679e4d1299f089ceca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2e38c6b054f47ea88f35f748423b7d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.569410] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.927119] env[63418]: DEBUG nova.compute.manager [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Received event network-changed-db5be14b-80a3-4322-b1a0-e6354d1eb8e4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 699.927119] env[63418]: DEBUG nova.compute.manager [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Refreshing instance network info cache due to event network-changed-db5be14b-80a3-4322-b1a0-e6354d1eb8e4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 699.927390] env[63418]: DEBUG oslo_concurrency.lockutils [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] Acquiring lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.937689] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 700.075706] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Releasing lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.076230] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 700.076428] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.076730] env[63418]: DEBUG oslo_concurrency.lockutils [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] Acquired lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.076898] env[63418]: DEBUG nova.network.neutron [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Refreshing network info cache for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 700.077902] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59fc58f0-71e2-4d73-916c-69d1758912e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.087909] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89d1dd1-4e1c-4c3d-9987-47a63ebdace2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.112016] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c78df16-faf1-4007-8492-bdc7cbfc608a could not be found. [ 700.112259] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.112429] env[63418]: INFO nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 700.112663] env[63418]: DEBUG oslo.service.loopingcall [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.115099] env[63418]: DEBUG nova.compute.manager [-] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 700.115275] env[63418]: DEBUG nova.network.neutron [-] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.147670] env[63418]: DEBUG nova.network.neutron [-] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.216119] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Successfully created port: 5576ba7a-3567-4754-aa5d-982a953254ee {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.418337] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a71c45-07f4-42ac-ab85-023388536490 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.427374] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48aa74ec-72a0-4463-9aba-85ccbcb1a5d3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.464144] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a415403-5664-414e-a78e-7e032ea03a21 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.475583] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29d58b7-f9e2-4e21-b7c8-42f81019ad64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.487389] env[63418]: DEBUG nova.compute.provider_tree [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.625767] env[63418]: DEBUG nova.network.neutron [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.653524] env[63418]: DEBUG nova.network.neutron [-] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.749521] env[63418]: DEBUG nova.network.neutron [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.968175] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 700.993735] env[63418]: DEBUG nova.scheduler.client.report [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 701.009927] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:38:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='cda3d2c9-0715-4b2f-b5ab-c41252461767',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-211379747',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.010433] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.010676] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.010878] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.012724] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.012724] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.012724] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.012724] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.012724] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.012949] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.012949] env[63418]: DEBUG nova.virt.hardware [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.013038] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c124e8c-12aa-462a-a1a8-568e6c6f4c5f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.022441] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770dbf2a-08d1-4c48-98ad-96cfef530164 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.156318] env[63418]: INFO nova.compute.manager [-] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Took 1.04 seconds to deallocate network for instance. [ 701.158942] env[63418]: DEBUG nova.compute.claims [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.159146] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.253477] env[63418]: DEBUG oslo_concurrency.lockutils [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] Releasing lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.253986] env[63418]: DEBUG nova.compute.manager [req-23282271-145f-4a62-8ef4-e4cf74dab46f req-0aa30a5b-9f6b-483b-a90c-1063211a5b05 service nova] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Received event network-vif-deleted-db5be14b-80a3-4322-b1a0-e6354d1eb8e4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 701.318782] env[63418]: DEBUG nova.compute.manager [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Received event network-changed-5576ba7a-3567-4754-aa5d-982a953254ee {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 701.319238] env[63418]: DEBUG nova.compute.manager [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Refreshing instance network info cache due to event network-changed-5576ba7a-3567-4754-aa5d-982a953254ee. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 701.319449] env[63418]: DEBUG oslo_concurrency.lockutils [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] Acquiring lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.320308] env[63418]: DEBUG oslo_concurrency.lockutils [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] Acquired lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.320308] env[63418]: DEBUG nova.network.neutron [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Refreshing network info cache for port 5576ba7a-3567-4754-aa5d-982a953254ee {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.421952] env[63418]: ERROR nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 701.421952] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.421952] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.421952] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.421952] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.421952] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.421952] env[63418]: ERROR nova.compute.manager raise self.value [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.421952] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.421952] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.421952] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.422940] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.422940] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.422940] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 701.422940] env[63418]: ERROR nova.compute.manager [ 701.422940] env[63418]: Traceback (most recent call last): [ 701.422940] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.422940] env[63418]: listener.cb(fileno) [ 701.422940] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.422940] env[63418]: result = function(*args, **kwargs) [ 701.422940] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.422940] env[63418]: return func(*args, **kwargs) [ 701.422940] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.422940] env[63418]: raise e [ 701.422940] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.422940] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 701.422940] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.422940] env[63418]: created_port_ids = self._update_ports_for_instance( [ 701.422940] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.422940] env[63418]: with excutils.save_and_reraise_exception(): [ 701.422940] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.422940] env[63418]: self.force_reraise() [ 701.422940] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.422940] env[63418]: raise self.value [ 701.422940] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.422940] env[63418]: updated_port = self._update_port( [ 701.422940] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.422940] env[63418]: _ensure_no_port_binding_failure(port) [ 701.422940] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.422940] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.423916] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 701.423916] env[63418]: Removing descriptor: 16 [ 701.423916] env[63418]: ERROR nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Traceback (most recent call last): [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] yield resources [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self.driver.spawn(context, instance, image_meta, [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.423916] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] vm_ref = self.build_virtual_machine(instance, [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] for vif in network_info: [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return self._sync_wrapper(fn, *args, **kwargs) [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self.wait() [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self[:] = self._gt.wait() [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return self._exit_event.wait() [ 701.424376] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] result = hub.switch() [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return self.greenlet.switch() [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] result = function(*args, **kwargs) [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return func(*args, **kwargs) [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] raise e [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] nwinfo = self.network_api.allocate_for_instance( [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.424763] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] created_port_ids = self._update_ports_for_instance( [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] with excutils.save_and_reraise_exception(): [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self.force_reraise() [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] raise self.value [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] updated_port = self._update_port( [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] _ensure_no_port_binding_failure(port) [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.425608] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] raise exception.PortBindingFailed(port_id=port['id']) [ 701.426226] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 701.426226] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] [ 701.426226] env[63418]: INFO nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Terminating instance [ 701.499778] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.500345] env[63418]: DEBUG nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 701.503796] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.905s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.838116] env[63418]: DEBUG nova.network.neutron [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.884996] env[63418]: DEBUG nova.network.neutron [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.929617] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.011281] env[63418]: DEBUG nova.compute.utils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.016869] env[63418]: DEBUG nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Not allocating networking since 'none' was specified. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 702.390547] env[63418]: DEBUG oslo_concurrency.lockutils [req-2f10e2bb-d994-43a7-90cc-c8ac235e4724 req-1bcf8fa6-1d9c-4d2a-ab79-3a2d58ea3877 service nova] Releasing lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.392024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquired lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.392241] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.424875] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28019f9-61e2-4984-89af-167d8e72a838 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.436914] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2291fe2a-ac12-4cb5-b9e5-fd46764a064f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.469997] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08cdf4d-5210-488c-bca4-420a24c02e44 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.481654] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d71549f-56c5-41a4-859e-38ae2175aecc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.494943] env[63418]: DEBUG nova.compute.provider_tree [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.518944] env[63418]: DEBUG nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 702.923541] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.972973] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.000798] env[63418]: DEBUG nova.scheduler.client.report [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 703.163369] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.163691] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.346386] env[63418]: DEBUG nova.compute.manager [req-b12102d8-477c-4b34-9f9c-6b0a876bad8a req-8d476360-afc8-4b18-8d8e-b4d9b6217be8 service nova] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Received event network-vif-deleted-5576ba7a-3567-4754-aa5d-982a953254ee {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 703.475947] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Releasing lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.476609] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 703.476913] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.477667] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8686da36-a114-42a6-8962-10cec1ab8f75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.486741] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4347255-30f5-4076-8b1c-234874980ba3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.509053] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.005s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.509531] env[63418]: ERROR nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Traceback (most recent call last): [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self.driver.spawn(context, instance, image_meta, [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] vm_ref = self.build_virtual_machine(instance, [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.509531] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] for vif in network_info: [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return self._sync_wrapper(fn, *args, **kwargs) [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self.wait() [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self[:] = self._gt.wait() [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return self._exit_event.wait() [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] result = hub.switch() [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.509877] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return self.greenlet.switch() [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] result = function(*args, **kwargs) [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] return func(*args, **kwargs) [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] raise e [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] nwinfo = self.network_api.allocate_for_instance( [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] created_port_ids = self._update_ports_for_instance( [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] with excutils.save_and_reraise_exception(): [ 703.510249] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] self.force_reraise() [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] raise self.value [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] updated_port = self._update_port( [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] _ensure_no_port_binding_failure(port) [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] raise exception.PortBindingFailed(port_id=port['id']) [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] nova.exception.PortBindingFailed: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. [ 703.510589] env[63418]: ERROR nova.compute.manager [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] [ 703.510915] env[63418]: DEBUG nova.compute.utils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.511583] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a80e0c1-2088-4de3-9032-d05ff48c3fcb could not be found. [ 703.511768] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.511936] env[63418]: INFO nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 703.512178] env[63418]: DEBUG oslo.service.loopingcall [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.512604] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Build of instance f6d62d1b-0333-40ac-92e8-a798e58d495c was re-scheduled: Binding failed for port 07a1d165-1209-42f0-b086-037ed171d4e4, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 703.512991] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 703.513217] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquiring lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.513418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Acquired lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.513586] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.514540] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.592s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.515987] env[63418]: INFO nova.compute.claims [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.518268] env[63418]: DEBUG nova.compute.manager [-] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 703.518368] env[63418]: DEBUG nova.network.neutron [-] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.527512] env[63418]: DEBUG nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 703.531909] env[63418]: DEBUG nova.network.neutron [-] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.552992] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.553252] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.553405] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.553633] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.553780] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.553922] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.554132] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.554288] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.554474] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.554616] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.554782] env[63418]: DEBUG nova.virt.hardware [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.555635] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02238ab3-2c22-444e-bb91-49e26231bf94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.563781] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fa7e24-2539-4fa7-83fd-fce91733b023 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.576920] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.582460] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Creating folder: Project (0a344237d3e243909f72f13d61235b33). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.583389] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6999cbbb-b055-4f28-936d-b14f7c3ddd6f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.592758] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Created folder: Project (0a344237d3e243909f72f13d61235b33) in parent group-v268354. [ 703.592941] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Creating folder: Instances. Parent ref: group-v268383. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.593401] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1835db53-ed1c-4a8e-9cb9-08c7b5c42c2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.601949] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Created folder: Instances in parent group-v268383. [ 703.602185] env[63418]: DEBUG oslo.service.loopingcall [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.602363] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 703.602545] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-966b5cb1-c51b-4e82-9ac7-94e6fd1035ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.618337] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 703.618337] env[63418]: value = "task-1244727" [ 703.618337] env[63418]: _type = "Task" [ 703.618337] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.626087] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244727, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.033897] env[63418]: DEBUG nova.network.neutron [-] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.039752] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.128152] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244727, 'name': CreateVM_Task, 'duration_secs': 0.242445} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.128317] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 704.128735] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.128891] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.129448] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 704.129519] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0af0bbd7-d653-4ad4-aa66-36e6f2289c84 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.134045] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 704.134045] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5292817c-12c5-548e-2dbc-2db3ee8ed313" [ 704.134045] env[63418]: _type = "Task" [ 704.134045] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.141647] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5292817c-12c5-548e-2dbc-2db3ee8ed313, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.171601] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.537306] env[63418]: INFO nova.compute.manager [-] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Took 1.02 seconds to deallocate network for instance. [ 704.539922] env[63418]: DEBUG nova.compute.claims [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.540222] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.648293] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5292817c-12c5-548e-2dbc-2db3ee8ed313, 'name': SearchDatastore_Task, 'duration_secs': 0.009417} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.648587] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.648818] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.649052] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.649234] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.649367] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.651780] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a91a73a0-a926-4b0f-b755-6b86b3f6f3c1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.660720] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.660914] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 704.661651] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ec1da0c-4b93-4200-a7a6-682146c8933d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.669063] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 704.669063] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bf73f0-bb3c-cea4-0586-7e2b2f8f0163" [ 704.669063] env[63418]: _type = "Task" [ 704.669063] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.676714] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bf73f0-bb3c-cea4-0586-7e2b2f8f0163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.679265] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Releasing lock "refresh_cache-f6d62d1b-0333-40ac-92e8-a798e58d495c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.679473] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 704.679648] env[63418]: DEBUG nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 704.679812] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.709846] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.896707] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6d1c97-6a8a-4eb8-ad21-cc6c1b5cd6ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.904049] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebd10d3-f18e-44d4-9e68-d2890bc178b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.933237] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d765a8a-19e5-4b8e-9243-72e7d9cc46e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.940956] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1fac90-3a72-4c1a-8152-eb6f26a2e666 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.954139] env[63418]: DEBUG nova.compute.provider_tree [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.179543] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bf73f0-bb3c-cea4-0586-7e2b2f8f0163, 'name': SearchDatastore_Task, 'duration_secs': 0.008947} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.180319] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70e52cec-c75a-46a0-867f-f9684abf8337 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.185054] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 705.185054] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ae3ee1-9e84-a53b-dcf8-db1d733b5bf1" [ 705.185054] env[63418]: _type = "Task" [ 705.185054] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.192343] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ae3ee1-9e84-a53b-dcf8-db1d733b5bf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.212946] env[63418]: DEBUG nova.network.neutron [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.457241] env[63418]: DEBUG nova.scheduler.client.report [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 705.695773] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ae3ee1-9e84-a53b-dcf8-db1d733b5bf1, 'name': SearchDatastore_Task, 'duration_secs': 0.008865} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.695921] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.696173] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.696426] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6d3283d-d693-4c13-9fb8-49f2d6e8e34e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.703433] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 705.703433] env[63418]: value = "task-1244728" [ 705.703433] env[63418]: _type = "Task" [ 705.703433] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.710640] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244728, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.715330] env[63418]: INFO nova.compute.manager [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] [instance: f6d62d1b-0333-40ac-92e8-a798e58d495c] Took 1.04 seconds to deallocate network for instance. [ 705.963580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.964180] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 705.967679] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.020s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.969195] env[63418]: INFO nova.compute.claims [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.213294] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244728, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457299} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.213646] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.213818] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.213986] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07275ac3-658f-43bc-b78e-ac64004c3777 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.223139] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 706.223139] env[63418]: value = "task-1244729" [ 706.223139] env[63418]: _type = "Task" [ 706.223139] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.230590] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244729, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.474066] env[63418]: DEBUG nova.compute.utils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.478067] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 706.478067] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 706.524681] env[63418]: DEBUG nova.policy [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ff7e8653f3f4fcab16db4b1154ee131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a88949fc1a184a6eb326f3bcf1fc0d49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.736150] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244729, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061804} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.736444] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 706.737198] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65367939-e8f8-4d39-8383-8e2bc845fb0b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.757161] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.758151] env[63418]: INFO nova.scheduler.client.report [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Deleted allocations for instance f6d62d1b-0333-40ac-92e8-a798e58d495c [ 706.764943] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-123c92ff-d075-45a0-a157-dd6ae85011aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.788600] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 706.788600] env[63418]: value = "task-1244730" [ 706.788600] env[63418]: _type = "Task" [ 706.788600] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.797720] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244730, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.978775] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 707.090642] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Successfully created port: 4eb0a23f-f609-4da1-bf11-a41617596ab5 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.266512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2f0f5b33-5e94-4937-b3f8-7f2aac4aefbc tempest-ServerAddressesTestJSON-1602281293 tempest-ServerAddressesTestJSON-1602281293-project-member] Lock "f6d62d1b-0333-40ac-92e8-a798e58d495c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.036s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.299568] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244730, 'name': ReconfigVM_Task, 'duration_secs': 0.498435} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.299849] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.300553] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67e1d626-bff2-43be-acbf-e6c62142a7c9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.306704] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 707.306704] env[63418]: value = "task-1244731" [ 707.306704] env[63418]: _type = "Task" [ 707.306704] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.317466] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244731, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.431428] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a81042e-0b99-4c27-8c2b-341d2f2545fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.438907] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff92087-29fe-4b4e-8f2e-9132ddc240a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.470216] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231a44f8-a86a-4f33-9a28-7d0f404b2323 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.477730] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4625fbc3-3517-49b9-8e67-556a78ff1eb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.494058] env[63418]: DEBUG nova.compute.provider_tree [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.768405] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 707.819509] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244731, 'name': Rename_Task, 'duration_secs': 0.13078} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.819775] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.820009] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-136647e6-cbeb-45d6-9a4b-405e87a95b8d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.825742] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 707.825742] env[63418]: value = "task-1244732" [ 707.825742] env[63418]: _type = "Task" [ 707.825742] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.834138] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244732, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.002067] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 708.003423] env[63418]: DEBUG nova.scheduler.client.report [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 708.048018] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.048018] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.048018] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.048226] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.048226] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.048403] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.048752] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.049073] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.049349] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.049614] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.049884] env[63418]: DEBUG nova.virt.hardware [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.051050] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711e25ff-f67a-4eca-9844-ad0bc89e0db5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.060158] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911a3fec-8810-43b9-ab5f-165e8b42735b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.148163] env[63418]: DEBUG nova.compute.manager [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Received event network-changed-4eb0a23f-f609-4da1-bf11-a41617596ab5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 708.148163] env[63418]: DEBUG nova.compute.manager [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Refreshing instance network info cache due to event network-changed-4eb0a23f-f609-4da1-bf11-a41617596ab5. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 708.148163] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] Acquiring lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.148163] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] Acquired lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.148163] env[63418]: DEBUG nova.network.neutron [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Refreshing network info cache for port 4eb0a23f-f609-4da1-bf11-a41617596ab5 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 708.297476] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.341144] env[63418]: DEBUG oslo_vmware.api [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244732, 'name': PowerOnVM_Task, 'duration_secs': 0.431764} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.341144] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.341144] env[63418]: INFO nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Took 4.81 seconds to spawn the instance on the hypervisor. [ 708.341540] env[63418]: DEBUG nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 708.342475] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4928ac5d-d624-4cd2-b081-56473716c777 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.373357] env[63418]: ERROR nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 708.373357] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 708.373357] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.373357] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.373357] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.373357] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.373357] env[63418]: ERROR nova.compute.manager raise self.value [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.373357] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.373357] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.373357] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.373928] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.373928] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.373928] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 708.373928] env[63418]: ERROR nova.compute.manager [ 708.373928] env[63418]: Traceback (most recent call last): [ 708.373928] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.373928] env[63418]: listener.cb(fileno) [ 708.373928] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.373928] env[63418]: result = function(*args, **kwargs) [ 708.373928] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.373928] env[63418]: return func(*args, **kwargs) [ 708.373928] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 708.373928] env[63418]: raise e [ 708.373928] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 708.373928] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 708.373928] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.373928] env[63418]: created_port_ids = self._update_ports_for_instance( [ 708.373928] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.373928] env[63418]: with excutils.save_and_reraise_exception(): [ 708.373928] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.373928] env[63418]: self.force_reraise() [ 708.373928] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.373928] env[63418]: raise self.value [ 708.373928] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.373928] env[63418]: updated_port = self._update_port( [ 708.373928] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.373928] env[63418]: _ensure_no_port_binding_failure(port) [ 708.373928] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.373928] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.375417] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 708.375417] env[63418]: Removing descriptor: 16 [ 708.375417] env[63418]: ERROR nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Traceback (most recent call last): [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] yield resources [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self.driver.spawn(context, instance, image_meta, [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.375417] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] vm_ref = self.build_virtual_machine(instance, [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] for vif in network_info: [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return self._sync_wrapper(fn, *args, **kwargs) [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self.wait() [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self[:] = self._gt.wait() [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return self._exit_event.wait() [ 708.375827] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] result = hub.switch() [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return self.greenlet.switch() [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] result = function(*args, **kwargs) [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return func(*args, **kwargs) [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] raise e [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] nwinfo = self.network_api.allocate_for_instance( [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.376241] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] created_port_ids = self._update_ports_for_instance( [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] with excutils.save_and_reraise_exception(): [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self.force_reraise() [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] raise self.value [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] updated_port = self._update_port( [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] _ensure_no_port_binding_failure(port) [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.376636] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] raise exception.PortBindingFailed(port_id=port['id']) [ 708.376969] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 708.376969] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] [ 708.376969] env[63418]: INFO nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Terminating instance [ 708.511102] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.511614] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 708.514387] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.860s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.680017] env[63418]: DEBUG nova.network.neutron [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.816998] env[63418]: DEBUG nova.network.neutron [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.867388] env[63418]: INFO nova.compute.manager [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Took 26.51 seconds to build instance. [ 708.885178] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquiring lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.019207] env[63418]: DEBUG nova.compute.utils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.023636] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 709.023636] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 709.079916] env[63418]: DEBUG nova.policy [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d6ceb0a847d4e95a026760b7cca68e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '372ef69dfb6143a9a504c39328afff9c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.321580] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e26206-1881-4a1e-a862-cf516ded14c3 req-9eb0ea16-a5eb-4f98-be78-ffb02c3f925e service nova] Releasing lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.321991] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquired lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.322326] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.369712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f18237e2-6caf-4ad2-ba65-1ad825d28b07 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.810s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.466750] env[63418]: INFO nova.compute.manager [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Rebuilding instance [ 709.471168] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d4940f-3680-4ca2-8558-709c9641f7a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.478544] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5013c023-3a1f-4888-801d-0877e5bca308 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.522108] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcb123a-38f3-4836-8959-dd50564534cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.529118] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 709.539096] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bd45c3-250e-4124-b596-443ebab55610 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.545426] env[63418]: DEBUG nova.compute.manager [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 709.546226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bcd195-7e8e-4d2a-add1-cdee2d933b8e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.557333] env[63418]: DEBUG nova.compute.provider_tree [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.700163] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Successfully created port: 50ca70cc-d061-4a2b-9244-01d875347dfd {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.842112] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.872803] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 709.923593] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.063026] env[63418]: DEBUG nova.scheduler.client.report [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 710.175583] env[63418]: DEBUG nova.compute.manager [req-5af3110a-f59e-410e-ab23-490e1dc67cf0 req-22bf4649-fa35-44b5-a238-c9d14449b8c3 service nova] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Received event network-vif-deleted-4eb0a23f-f609-4da1-bf11-a41617596ab5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 710.396045] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.426751] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Releasing lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.427199] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 710.427393] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 710.427706] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4bcf2fe-e816-4467-bdf7-a195cfb5ce46 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.436812] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be35cf1d-1c3f-4b01-a7ac-d6f7bee8e653 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.459365] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c79aa3d1-bbca-45cf-9f29-40a2720c4437 could not be found. [ 710.459638] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.459960] env[63418]: INFO nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Took 0.03 seconds to destroy the instance on the hypervisor. [ 710.460264] env[63418]: DEBUG oslo.service.loopingcall [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.460535] env[63418]: DEBUG nova.compute.manager [-] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 710.460663] env[63418]: DEBUG nova.network.neutron [-] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.480940] env[63418]: DEBUG nova.network.neutron [-] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.542369] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 710.567118] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.052s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.567118] env[63418]: ERROR nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Traceback (most recent call last): [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self.driver.spawn(context, instance, image_meta, [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.567118] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] vm_ref = self.build_virtual_machine(instance, [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] for vif in network_info: [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return self._sync_wrapper(fn, *args, **kwargs) [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self.wait() [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self[:] = self._gt.wait() [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return self._exit_event.wait() [ 710.567797] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] result = hub.switch() [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return self.greenlet.switch() [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] result = function(*args, **kwargs) [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] return func(*args, **kwargs) [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] raise e [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] nwinfo = self.network_api.allocate_for_instance( [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.568212] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] created_port_ids = self._update_ports_for_instance( [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] with excutils.save_and_reraise_exception(): [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] self.force_reraise() [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] raise self.value [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] updated_port = self._update_port( [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] _ensure_no_port_binding_failure(port) [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.568657] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] raise exception.PortBindingFailed(port_id=port['id']) [ 710.569040] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] nova.exception.PortBindingFailed: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. [ 710.569040] env[63418]: ERROR nova.compute.manager [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] [ 710.570048] env[63418]: DEBUG nova.compute.utils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.573808] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.574056] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.574214] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.574392] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.574535] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.574705] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.574970] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.575145] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.575311] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.575470] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.575636] env[63418]: DEBUG nova.virt.hardware [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.575972] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.111s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.579156] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 710.579960] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b782d02a-5d2a-4f91-bb3e-b3f7fd30332b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.583331] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Build of instance 8e39c8ef-103a-45f4-8630-bfd51b38445d was re-scheduled: Binding failed for port 75df9a9b-8bf5-414c-99c5-f3897ca7e448, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 710.583659] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 710.583882] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquiring lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.584336] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Acquired lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.584336] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.585178] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9175a02-2b26-4178-aabd-159b4269babe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.594214] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392cdcc9-8651-460e-bba5-04cf9e1154ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.598145] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 710.598145] env[63418]: value = "task-1244733" [ 710.598145] env[63418]: _type = "Task" [ 710.598145] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.616370] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.891987] env[63418]: ERROR nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 710.891987] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 710.891987] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.891987] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.891987] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.891987] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.891987] env[63418]: ERROR nova.compute.manager raise self.value [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.891987] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.891987] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.891987] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.892581] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.892581] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.892581] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 710.892581] env[63418]: ERROR nova.compute.manager [ 710.892581] env[63418]: Traceback (most recent call last): [ 710.892581] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.892581] env[63418]: listener.cb(fileno) [ 710.892581] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.892581] env[63418]: result = function(*args, **kwargs) [ 710.892581] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.892581] env[63418]: return func(*args, **kwargs) [ 710.892581] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 710.892581] env[63418]: raise e [ 710.892581] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 710.892581] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 710.892581] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.892581] env[63418]: created_port_ids = self._update_ports_for_instance( [ 710.892581] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.892581] env[63418]: with excutils.save_and_reraise_exception(): [ 710.892581] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.892581] env[63418]: self.force_reraise() [ 710.892581] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.892581] env[63418]: raise self.value [ 710.892581] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.892581] env[63418]: updated_port = self._update_port( [ 710.892581] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.892581] env[63418]: _ensure_no_port_binding_failure(port) [ 710.892581] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.892581] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.893355] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 710.893355] env[63418]: Removing descriptor: 16 [ 710.893355] env[63418]: ERROR nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] Traceback (most recent call last): [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] yield resources [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self.driver.spawn(context, instance, image_meta, [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.893355] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] vm_ref = self.build_virtual_machine(instance, [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] for vif in network_info: [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return self._sync_wrapper(fn, *args, **kwargs) [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self.wait() [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self[:] = self._gt.wait() [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return self._exit_event.wait() [ 710.893751] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] result = hub.switch() [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return self.greenlet.switch() [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] result = function(*args, **kwargs) [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return func(*args, **kwargs) [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] raise e [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] nwinfo = self.network_api.allocate_for_instance( [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.894125] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] created_port_ids = self._update_ports_for_instance( [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] with excutils.save_and_reraise_exception(): [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self.force_reraise() [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] raise self.value [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] updated_port = self._update_port( [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] _ensure_no_port_binding_failure(port) [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.894498] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] raise exception.PortBindingFailed(port_id=port['id']) [ 710.894826] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 710.894826] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] [ 710.894826] env[63418]: INFO nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Terminating instance [ 710.984042] env[63418]: DEBUG nova.network.neutron [-] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.103881] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.110800] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244733, 'name': PowerOffVM_Task, 'duration_secs': 0.186978} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.113285] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 711.113545] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.114586] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c19fa79-f8b9-443e-8abf-6a99dc0fdf2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.121063] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 711.121303] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44a28ba8-8cbb-4742-ad6f-5b6241903226 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.146498] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 711.146707] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 711.146879] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Deleting the datastore file [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 711.147159] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20ce1b34-f917-472d-972c-ac1ad93fe8b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.153849] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 711.153849] env[63418]: value = "task-1244735" [ 711.153849] env[63418]: _type = "Task" [ 711.153849] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.164934] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244735, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.192987] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.400722] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.401029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquired lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.401029] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.433245] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bbe5bf-71e8-4600-8936-461cd61ebaf9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.441824] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a359ee-6e17-4f48-9e4a-4e4e86d1c157 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.472447] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a953af-929c-4af0-9740-d37c23bac573 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.480318] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddce9ae-6b32-4f67-aa15-a62401ed1f13 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.494267] env[63418]: INFO nova.compute.manager [-] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Took 1.03 seconds to deallocate network for instance. [ 711.494747] env[63418]: DEBUG nova.compute.provider_tree [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.499849] env[63418]: DEBUG nova.compute.claims [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.499937] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.664348] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244735, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102893} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.664648] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 711.664875] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 711.665095] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.696046] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Releasing lock "refresh_cache-8e39c8ef-103a-45f4-8630-bfd51b38445d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.696476] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 711.696476] env[63418]: DEBUG nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 711.696684] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.718733] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.917272] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.998416] env[63418]: DEBUG nova.scheduler.client.report [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 712.024988] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.206935] env[63418]: DEBUG nova.compute.manager [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] [instance: 96183939-654d-4510-adf2-303f41f94640] Received event network-changed-50ca70cc-d061-4a2b-9244-01d875347dfd {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 712.207174] env[63418]: DEBUG nova.compute.manager [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] [instance: 96183939-654d-4510-adf2-303f41f94640] Refreshing instance network info cache due to event network-changed-50ca70cc-d061-4a2b-9244-01d875347dfd. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 712.207348] env[63418]: DEBUG oslo_concurrency.lockutils [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] Acquiring lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.220938] env[63418]: DEBUG nova.network.neutron [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.503046] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.503698] env[63418]: ERROR nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Traceback (most recent call last): [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self.driver.spawn(context, instance, image_meta, [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] vm_ref = self.build_virtual_machine(instance, [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.503698] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] for vif in network_info: [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] return self._sync_wrapper(fn, *args, **kwargs) [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self.wait() [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self[:] = self._gt.wait() [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] return self._exit_event.wait() [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] current.throw(*self._exc) [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.504110] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] result = function(*args, **kwargs) [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] return func(*args, **kwargs) [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] raise e [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] nwinfo = self.network_api.allocate_for_instance( [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] created_port_ids = self._update_ports_for_instance( [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] with excutils.save_and_reraise_exception(): [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] self.force_reraise() [ 712.504524] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] raise self.value [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] updated_port = self._update_port( [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] _ensure_no_port_binding_failure(port) [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] raise exception.PortBindingFailed(port_id=port['id']) [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] nova.exception.PortBindingFailed: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. [ 712.504952] env[63418]: ERROR nova.compute.manager [instance: 69651fab-08a7-4832-8c91-2fa01d186860] [ 712.504952] env[63418]: DEBUG nova.compute.utils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.505751] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.424s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.507221] env[63418]: INFO nova.compute.claims [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.509981] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Build of instance 69651fab-08a7-4832-8c91-2fa01d186860 was re-scheduled: Binding failed for port 1295a3e6-1a03-49d8-bfaa-fe456210135f, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 712.510382] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 712.510757] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.510757] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.510871] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.527294] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Releasing lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.527683] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 712.527920] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.528401] env[63418]: DEBUG oslo_concurrency.lockutils [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] Acquired lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.528580] env[63418]: DEBUG nova.network.neutron [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] [instance: 96183939-654d-4510-adf2-303f41f94640] Refreshing network info cache for port 50ca70cc-d061-4a2b-9244-01d875347dfd {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 712.529458] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a6a0400-bb49-42b6-bdeb-6a299f8ee766 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.539634] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88327ce2-ca41-4e60-ba8a-f3539600d243 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.564073] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96183939-654d-4510-adf2-303f41f94640 could not be found. [ 712.564220] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.564337] env[63418]: INFO nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Took 0.04 seconds to destroy the instance on the hypervisor. [ 712.564649] env[63418]: DEBUG oslo.service.loopingcall [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.564872] env[63418]: DEBUG nova.compute.manager [-] [instance: 96183939-654d-4510-adf2-303f41f94640] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 712.564962] env[63418]: DEBUG nova.network.neutron [-] [instance: 96183939-654d-4510-adf2-303f41f94640] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.582965] env[63418]: DEBUG nova.network.neutron [-] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.698210] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.698528] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.698688] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.698883] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.699040] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.699190] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.699394] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.699549] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.699715] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.699879] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.700063] env[63418]: DEBUG nova.virt.hardware [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.700895] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf53b83-fe4f-4399-9ac0-a9d9ef211a1d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.708480] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528e903a-7323-4a81-ba6d-ff06a595f24f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.724879] env[63418]: INFO nova.compute.manager [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] [instance: 8e39c8ef-103a-45f4-8630-bfd51b38445d] Took 1.03 seconds to deallocate network for instance. [ 712.727609] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance VIF info [] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 712.733171] env[63418]: DEBUG oslo.service.loopingcall [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.733603] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 712.733807] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-440a4039-466c-4c63-b387-6f90e2072327 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.749673] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 712.749673] env[63418]: value = "task-1244736" [ 712.749673] env[63418]: _type = "Task" [ 712.749673] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.759175] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244736, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.029641] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.047548] env[63418]: DEBUG nova.network.neutron [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.087049] env[63418]: DEBUG nova.network.neutron [-] [instance: 96183939-654d-4510-adf2-303f41f94640] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.130958] env[63418]: DEBUG nova.network.neutron [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] [instance: 96183939-654d-4510-adf2-303f41f94640] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.154910] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.259961] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244736, 'name': CreateVM_Task, 'duration_secs': 0.239625} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.260155] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 713.260568] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.260723] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.261059] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 713.261310] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-817c6764-4b2f-48cd-a5f9-af7dc0c38f1b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.265967] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 713.265967] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521ac081-a36c-e7e8-4d5e-f079c76310c6" [ 713.265967] env[63418]: _type = "Task" [ 713.265967] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.275118] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521ac081-a36c-e7e8-4d5e-f079c76310c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.593200] env[63418]: INFO nova.compute.manager [-] [instance: 96183939-654d-4510-adf2-303f41f94640] Took 1.03 seconds to deallocate network for instance. [ 713.595408] env[63418]: DEBUG nova.compute.claims [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 713.595583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.633520] env[63418]: DEBUG oslo_concurrency.lockutils [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] Releasing lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.633797] env[63418]: DEBUG nova.compute.manager [req-68792d57-ed58-4971-a2a8-1605670a3d3b req-7e12ba4a-816a-40e4-801b-c971ff42ecfd service nova] [instance: 96183939-654d-4510-adf2-303f41f94640] Received event network-vif-deleted-50ca70cc-d061-4a2b-9244-01d875347dfd {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 713.657045] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Releasing lock "refresh_cache-69651fab-08a7-4832-8c91-2fa01d186860" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.657293] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 713.657469] env[63418]: DEBUG nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 713.657634] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.764080] env[63418]: INFO nova.scheduler.client.report [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Deleted allocations for instance 8e39c8ef-103a-45f4-8630-bfd51b38445d [ 713.784579] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521ac081-a36c-e7e8-4d5e-f079c76310c6, 'name': SearchDatastore_Task, 'duration_secs': 0.008793} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.784888] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.785121] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.785352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.785498] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.785675] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.786273] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-124821b7-cb42-4429-bbdf-05a6ad208419 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.794619] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.794788] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 713.795500] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bed89983-7808-4721-96c1-16182d71cbbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.804870] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 713.804870] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cba38d-5ae5-a3f5-017e-20092301b6ab" [ 713.804870] env[63418]: _type = "Task" [ 713.804870] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.813107] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cba38d-5ae5-a3f5-017e-20092301b6ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.841868] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.909826] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987bdbc5-b6b8-4fe5-9762-8c86eb1ae5c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.917911] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4af8520-52aa-4f6c-9b2c-4493e688992a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.948417] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9475020-e459-4566-b9b1-c41a967ed1ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.955987] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b8fd8b-f7c7-49fc-a73d-983b0a226f64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.969172] env[63418]: DEBUG nova.compute.provider_tree [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.279126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b37c67c-2f99-4830-a53c-2f693fc27abb tempest-InstanceActionsTestJSON-787966284 tempest-InstanceActionsTestJSON-787966284-project-member] Lock "8e39c8ef-103a-45f4-8630-bfd51b38445d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.642s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.314502] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cba38d-5ae5-a3f5-017e-20092301b6ab, 'name': SearchDatastore_Task, 'duration_secs': 0.007591} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.315491] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-590eaab5-7633-4a4f-8cbc-6cce3904ae1b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.320451] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 714.320451] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5251c56b-df53-5f17-fb6c-9e020bba9893" [ 714.320451] env[63418]: _type = "Task" [ 714.320451] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.328377] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5251c56b-df53-5f17-fb6c-9e020bba9893, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.344964] env[63418]: DEBUG nova.network.neutron [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.472239] env[63418]: DEBUG nova.scheduler.client.report [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 714.781278] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 714.831038] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5251c56b-df53-5f17-fb6c-9e020bba9893, 'name': SearchDatastore_Task, 'duration_secs': 0.008994} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.831305] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.831556] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 714.831800] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-172cf12b-1505-45c8-985a-615d3b368b6a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.839582] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 714.839582] env[63418]: value = "task-1244737" [ 714.839582] env[63418]: _type = "Task" [ 714.839582] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.847342] env[63418]: INFO nova.compute.manager [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: 69651fab-08a7-4832-8c91-2fa01d186860] Took 1.19 seconds to deallocate network for instance. [ 714.849835] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.976870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.977465] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 714.980228] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.680s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.306308] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.349629] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244737, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465688} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.349811] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 715.349988] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 715.353340] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b9ca09d-85d6-41de-9f56-db971ef1ef77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.361384] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 715.361384] env[63418]: value = "task-1244738" [ 715.361384] env[63418]: _type = "Task" [ 715.361384] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.368978] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.486289] env[63418]: DEBUG nova.compute.utils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.493038] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 715.493038] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.585752] env[63418]: DEBUG nova.policy [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b726500666f241d49cd1edd63dc93d1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd90d339ec08c43bb909125ba3058b3ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.870238] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070706} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.872884] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 715.874046] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3622e2bd-a695-494e-81db-e59c6951eb64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.877161] env[63418]: INFO nova.scheduler.client.report [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Deleted allocations for instance 69651fab-08a7-4832-8c91-2fa01d186860 [ 715.905127] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 715.906927] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c574b4d5-2be8-4768-8c36-a7dffdaed53f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.921360] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc4be38-1b4c-4a66-b574-d7c42f5059d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.935138] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f53304a-e39b-49bb-938c-d723297b2a27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.935138] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 715.935138] env[63418]: value = "task-1244739" [ 715.935138] env[63418]: _type = "Task" [ 715.935138] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.966035] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96e5458-5aa8-4362-9a13-b7f80bc9b303 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.971701] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.976484] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2ef5cb-15ca-4184-9c1d-ff598d3fad5d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.989781] env[63418]: DEBUG nova.compute.provider_tree [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.992709] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 716.069208] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Successfully created port: 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.387549] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9853561-376e-43c7-85ef-2e169a7b0be7 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "69651fab-08a7-4832-8c91-2fa01d186860" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.054s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.444248] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244739, 'name': ReconfigVM_Task, 'duration_secs': 0.258414} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.444538] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b/99d00df2-6bf3-4ffe-b77d-f44ada631c8b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 716.445217] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33652869-bc50-40f8-a49a-8e071632379c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.451695] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 716.451695] env[63418]: value = "task-1244740" [ 716.451695] env[63418]: _type = "Task" [ 716.451695] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.460285] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244740, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.492859] env[63418]: DEBUG nova.scheduler.client.report [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 716.890468] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 716.961167] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244740, 'name': Rename_Task, 'duration_secs': 0.136606} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.961450] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 716.961687] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9246550-bf86-4ada-9b07-bad4e2490a3c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.968874] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 716.968874] env[63418]: value = "task-1244741" [ 716.968874] env[63418]: _type = "Task" [ 716.968874] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.976228] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.000737] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.001466] env[63418]: ERROR nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Traceback (most recent call last): [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self.driver.spawn(context, instance, image_meta, [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] vm_ref = self.build_virtual_machine(instance, [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.001466] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] for vif in network_info: [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] return self._sync_wrapper(fn, *args, **kwargs) [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self.wait() [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self[:] = self._gt.wait() [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] return self._exit_event.wait() [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] current.throw(*self._exc) [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.002092] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] result = function(*args, **kwargs) [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] return func(*args, **kwargs) [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] raise e [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] nwinfo = self.network_api.allocate_for_instance( [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] created_port_ids = self._update_ports_for_instance( [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] with excutils.save_and_reraise_exception(): [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] self.force_reraise() [ 717.002452] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] raise self.value [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] updated_port = self._update_port( [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] _ensure_no_port_binding_failure(port) [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] raise exception.PortBindingFailed(port_id=port['id']) [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] nova.exception.PortBindingFailed: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. [ 717.002808] env[63418]: ERROR nova.compute.manager [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] [ 717.002808] env[63418]: DEBUG nova.compute.utils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.004625] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.845s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.011024] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 717.011777] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Build of instance 2edbeef4-be3b-4a1d-8895-8410b71b6b18 was re-scheduled: Binding failed for port c6429985-1dff-4d47-967a-a20c8738a942, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 717.012246] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 717.015016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquiring lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.015016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Acquired lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.015016] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.017524] env[63418]: DEBUG nova.compute.manager [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Received event network-changed-47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 717.017524] env[63418]: DEBUG nova.compute.manager [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Refreshing instance network info cache due to event network-changed-47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 717.017692] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] Acquiring lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.017756] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] Acquired lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.017872] env[63418]: DEBUG nova.network.neutron [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Refreshing network info cache for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.046632] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.046632] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.046632] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.046832] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.046832] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.047268] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.047791] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.051279] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.051279] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.051279] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.051279] env[63418]: DEBUG nova.virt.hardware [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.051279] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4e05b9-c171-4162-828a-a3e0e9f7bcce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.059404] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2054bf8-0d11-4fc2-a222-74e43f3a0de9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.255766] env[63418]: ERROR nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 717.255766] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 717.255766] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.255766] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.255766] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.255766] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.255766] env[63418]: ERROR nova.compute.manager raise self.value [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.255766] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.255766] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.255766] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.256418] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.256418] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.256418] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 717.256418] env[63418]: ERROR nova.compute.manager [ 717.256418] env[63418]: Traceback (most recent call last): [ 717.256418] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.256418] env[63418]: listener.cb(fileno) [ 717.256418] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.256418] env[63418]: result = function(*args, **kwargs) [ 717.256418] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.256418] env[63418]: return func(*args, **kwargs) [ 717.256418] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 717.256418] env[63418]: raise e [ 717.256418] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 717.256418] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 717.256418] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.256418] env[63418]: created_port_ids = self._update_ports_for_instance( [ 717.256418] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.256418] env[63418]: with excutils.save_and_reraise_exception(): [ 717.256418] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.256418] env[63418]: self.force_reraise() [ 717.256418] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.256418] env[63418]: raise self.value [ 717.256418] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.256418] env[63418]: updated_port = self._update_port( [ 717.256418] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.256418] env[63418]: _ensure_no_port_binding_failure(port) [ 717.256418] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.256418] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.257476] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 717.257476] env[63418]: Removing descriptor: 16 [ 717.257476] env[63418]: ERROR nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Traceback (most recent call last): [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] yield resources [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self.driver.spawn(context, instance, image_meta, [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.257476] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] vm_ref = self.build_virtual_machine(instance, [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] for vif in network_info: [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return self._sync_wrapper(fn, *args, **kwargs) [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self.wait() [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self[:] = self._gt.wait() [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return self._exit_event.wait() [ 717.257950] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] result = hub.switch() [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return self.greenlet.switch() [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] result = function(*args, **kwargs) [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return func(*args, **kwargs) [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] raise e [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] nwinfo = self.network_api.allocate_for_instance( [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.258505] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] created_port_ids = self._update_ports_for_instance( [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] with excutils.save_and_reraise_exception(): [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self.force_reraise() [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] raise self.value [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] updated_port = self._update_port( [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] _ensure_no_port_binding_failure(port) [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.259026] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] raise exception.PortBindingFailed(port_id=port['id']) [ 717.259374] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 717.259374] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] [ 717.259374] env[63418]: INFO nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Terminating instance [ 717.421933] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.478984] env[63418]: DEBUG oslo_vmware.api [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244741, 'name': PowerOnVM_Task, 'duration_secs': 0.467688} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.479283] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 717.479485] env[63418]: DEBUG nova.compute.manager [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 717.480365] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d45eeb-af65-4ff6-b6dc-bf6ca5219f2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.539668] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.541486] env[63418]: DEBUG nova.network.neutron [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.631781] env[63418]: DEBUG nova.network.neutron [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.706743] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.762037] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquiring lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.872128] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c841666-6670-45d5-a661-d3fd080d10e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.879409] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7062a405-ec33-46b9-b32d-f8c65947bf6c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.910169] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a527ede7-2e59-4e07-a7bc-01330b4d96a8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.917657] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10888432-07f5-4cad-a2bc-d00719c29a01 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.930521] env[63418]: DEBUG nova.compute.provider_tree [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.004104] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.138438] env[63418]: DEBUG oslo_concurrency.lockutils [req-98e1ecd1-55cb-4a0a-b02d-561c36542dc6 req-b5f04200-d7f6-4c39-8c0e-74e7692cd9e8 service nova] Releasing lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.138881] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquired lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.139093] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.211173] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Releasing lock "refresh_cache-2edbeef4-be3b-4a1d-8895-8410b71b6b18" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.211385] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 718.211593] env[63418]: DEBUG nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 718.211761] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.234093] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.350607] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.350898] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.351133] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.351319] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.351488] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.358159] env[63418]: INFO nova.compute.manager [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Terminating instance [ 718.434858] env[63418]: DEBUG nova.scheduler.client.report [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 718.659061] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.736148] env[63418]: DEBUG nova.network.neutron [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.792337] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.862097] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "refresh_cache-99d00df2-6bf3-4ffe-b77d-f44ada631c8b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.862282] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquired lock "refresh_cache-99d00df2-6bf3-4ffe-b77d-f44ada631c8b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.862456] env[63418]: DEBUG nova.network.neutron [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.939250] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.940219] env[63418]: ERROR nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Traceback (most recent call last): [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self.driver.spawn(context, instance, image_meta, [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] vm_ref = self.build_virtual_machine(instance, [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.940219] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] for vif in network_info: [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] return self._sync_wrapper(fn, *args, **kwargs) [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self.wait() [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self[:] = self._gt.wait() [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] return self._exit_event.wait() [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] current.throw(*self._exc) [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.940995] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] result = function(*args, **kwargs) [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] return func(*args, **kwargs) [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] raise e [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] nwinfo = self.network_api.allocate_for_instance( [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] created_port_ids = self._update_ports_for_instance( [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] with excutils.save_and_reraise_exception(): [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] self.force_reraise() [ 718.941702] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] raise self.value [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] updated_port = self._update_port( [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] _ensure_no_port_binding_failure(port) [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] raise exception.PortBindingFailed(port_id=port['id']) [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] nova.exception.PortBindingFailed: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. [ 718.942354] env[63418]: ERROR nova.compute.manager [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] [ 718.942354] env[63418]: DEBUG nova.compute.utils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.944403] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.403s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.947390] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Build of instance 8c78df16-faf1-4007-8492-bdc7cbfc608a was re-scheduled: Binding failed for port db5be14b-80a3-4322-b1a0-e6354d1eb8e4, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 718.947890] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 718.948419] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquiring lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.948574] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Acquired lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.948868] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.083051] env[63418]: DEBUG nova.compute.manager [req-eca9c979-fae2-428e-a1f4-c6789a8c55e3 req-40914147-94df-4ed3-937b-d21985930383 service nova] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Received event network-vif-deleted-47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 719.239635] env[63418]: INFO nova.compute.manager [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] [instance: 2edbeef4-be3b-4a1d-8895-8410b71b6b18] Took 1.03 seconds to deallocate network for instance. [ 719.295119] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Releasing lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.295507] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 719.295697] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.296151] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ab573d9-23da-429b-96a1-b184ba167eb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.306097] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6760694e-51c0-43aa-96fb-06e7b076c988 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.330530] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3 could not be found. [ 719.330912] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.331225] env[63418]: INFO nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 719.331626] env[63418]: DEBUG oslo.service.loopingcall [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.331975] env[63418]: DEBUG nova.compute.manager [-] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 719.332195] env[63418]: DEBUG nova.network.neutron [-] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.360928] env[63418]: DEBUG nova.network.neutron [-] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.384324] env[63418]: DEBUG nova.network.neutron [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.480174] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.484641] env[63418]: DEBUG nova.network.neutron [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.622910] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.864022] env[63418]: DEBUG nova.network.neutron [-] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.902387] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44455674-76af-4a92-a2ef-4062bd31481a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.910646] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff97fa8-451a-4f9c-a45c-299568e62c9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.944548] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1bf5b1-1de7-4f27-8c41-3be5e6da6a69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.952755] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8818031f-b0f0-451d-b631-10521d134554 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.968255] env[63418]: DEBUG nova.compute.provider_tree [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.989671] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Releasing lock "refresh_cache-99d00df2-6bf3-4ffe-b77d-f44ada631c8b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.990428] env[63418]: DEBUG nova.compute.manager [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 719.990849] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.991574] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afea412-dd10-4d47-a2ca-adee37f73c90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.998883] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 719.999147] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8df9c19-a534-4bac-ac53-614ef2d6f1a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.005527] env[63418]: DEBUG oslo_vmware.api [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 720.005527] env[63418]: value = "task-1244742" [ 720.005527] env[63418]: _type = "Task" [ 720.005527] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.013480] env[63418]: DEBUG oslo_vmware.api [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.126120] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Releasing lock "refresh_cache-8c78df16-faf1-4007-8492-bdc7cbfc608a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.126566] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 720.126566] env[63418]: DEBUG nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 720.126713] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.145545] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.274826] env[63418]: INFO nova.scheduler.client.report [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Deleted allocations for instance 2edbeef4-be3b-4a1d-8895-8410b71b6b18 [ 720.366943] env[63418]: INFO nova.compute.manager [-] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Took 1.03 seconds to deallocate network for instance. [ 720.369434] env[63418]: DEBUG nova.compute.claims [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 720.369611] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.473173] env[63418]: DEBUG nova.scheduler.client.report [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 720.515808] env[63418]: DEBUG oslo_vmware.api [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244742, 'name': PowerOffVM_Task, 'duration_secs': 0.287518} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.516078] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 720.516258] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 720.516514] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd00e3ec-5aaf-45d4-9368-80ba023ed714 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.539159] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 720.539373] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 720.539551] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Deleting the datastore file [datastore2] 99d00df2-6bf3-4ffe-b77d-f44ada631c8b {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 720.539809] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c246ba9a-3a00-434c-848f-6d97a2fc56f1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.546364] env[63418]: DEBUG oslo_vmware.api [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for the task: (returnval){ [ 720.546364] env[63418]: value = "task-1244744" [ 720.546364] env[63418]: _type = "Task" [ 720.546364] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.554431] env[63418]: DEBUG oslo_vmware.api [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.647990] env[63418]: DEBUG nova.network.neutron [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.784400] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e7b59f05-6162-4176-98bc-273621d63306 tempest-ServersTestMultiNic-715475192 tempest-ServersTestMultiNic-715475192-project-member] Lock "2edbeef4-be3b-4a1d-8895-8410b71b6b18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.104s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.926486] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.926486] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.977090] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.977684] env[63418]: ERROR nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Traceback (most recent call last): [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self.driver.spawn(context, instance, image_meta, [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] vm_ref = self.build_virtual_machine(instance, [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.977684] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] for vif in network_info: [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return self._sync_wrapper(fn, *args, **kwargs) [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self.wait() [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self[:] = self._gt.wait() [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return self._exit_event.wait() [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] result = hub.switch() [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.978148] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return self.greenlet.switch() [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] result = function(*args, **kwargs) [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] return func(*args, **kwargs) [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] raise e [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] nwinfo = self.network_api.allocate_for_instance( [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] created_port_ids = self._update_ports_for_instance( [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] with excutils.save_and_reraise_exception(): [ 720.978618] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] self.force_reraise() [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] raise self.value [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] updated_port = self._update_port( [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] _ensure_no_port_binding_failure(port) [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] raise exception.PortBindingFailed(port_id=port['id']) [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] nova.exception.PortBindingFailed: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. [ 720.979109] env[63418]: ERROR nova.compute.manager [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] [ 720.979509] env[63418]: DEBUG nova.compute.utils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.979546] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.682s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.980985] env[63418]: INFO nova.compute.claims [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.983494] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Build of instance 6a80e0c1-2088-4de3-9032-d05ff48c3fcb was re-scheduled: Binding failed for port 5576ba7a-3567-4754-aa5d-982a953254ee, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 720.983980] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 720.984210] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquiring lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.984355] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Acquired lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.984532] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.057246] env[63418]: DEBUG oslo_vmware.api [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Task: {'id': task-1244744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1599} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.057507] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 721.057687] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 721.057860] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.058038] env[63418]: INFO nova.compute.manager [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Took 1.07 seconds to destroy the instance on the hypervisor. [ 721.058513] env[63418]: DEBUG oslo.service.loopingcall [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.058513] env[63418]: DEBUG nova.compute.manager [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 721.058591] env[63418]: DEBUG nova.network.neutron [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.075125] env[63418]: DEBUG nova.network.neutron [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.150349] env[63418]: INFO nova.compute.manager [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] [instance: 8c78df16-faf1-4007-8492-bdc7cbfc608a] Took 1.02 seconds to deallocate network for instance. [ 721.287264] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 721.517120] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.577508] env[63418]: DEBUG nova.network.neutron [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.602940] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.809929] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.081918] env[63418]: INFO nova.compute.manager [-] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Took 1.02 seconds to deallocate network for instance. [ 722.105485] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Releasing lock "refresh_cache-6a80e0c1-2088-4de3-9032-d05ff48c3fcb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.105715] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 722.105879] env[63418]: DEBUG nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 722.106894] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.128405] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.180239] env[63418]: INFO nova.scheduler.client.report [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Deleted allocations for instance 8c78df16-faf1-4007-8492-bdc7cbfc608a [ 722.359358] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b10f300-d888-45c5-803b-535bf5979cf4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.368723] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fc7136-7ad7-4c42-8fe0-2f6800680524 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.405251] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1defd6-3b10-4eac-9ae4-437d38c12e49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.413203] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20837d63-7c55-4b40-b958-fa57e0903b43 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.426441] env[63418]: DEBUG nova.compute.provider_tree [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.595793] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.633399] env[63418]: DEBUG nova.network.neutron [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.690473] env[63418]: DEBUG oslo_concurrency.lockutils [None req-35ca53c1-81d1-4348-873a-3156c3d18450 tempest-ServerActionsTestJSON-689494756 tempest-ServerActionsTestJSON-689494756-project-member] Lock "8c78df16-faf1-4007-8492-bdc7cbfc608a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.675s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.931270] env[63418]: DEBUG nova.scheduler.client.report [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 723.101369] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.101369] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.135884] env[63418]: INFO nova.compute.manager [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] [instance: 6a80e0c1-2088-4de3-9032-d05ff48c3fcb] Took 1.03 seconds to deallocate network for instance. [ 723.194969] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 723.435618] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.436343] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 723.439566] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.044s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.441365] env[63418]: INFO nova.compute.claims [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.608401] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.608401] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 723.727483] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.951348] env[63418]: DEBUG nova.compute.utils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.951348] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 723.951348] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.026161] env[63418]: DEBUG nova.policy [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d6ceb0a847d4e95a026760b7cca68e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '372ef69dfb6143a9a504c39328afff9c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.168849] env[63418]: INFO nova.scheduler.client.report [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Deleted allocations for instance 6a80e0c1-2088-4de3-9032-d05ff48c3fcb [ 724.458204] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 724.480221] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Successfully created port: 746f56d6-c17d-4ebd-abcd-da41b3cc1855 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.619389] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Didn't find any instances for network info cache update. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10364}} [ 724.619389] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.619389] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.619389] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.619389] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.619389] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.619607] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.619607] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 724.619607] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.673236] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c1042311-b704-4805-9e10-cdd244933129 tempest-MigrationsAdminTest-799535964 tempest-MigrationsAdminTest-799535964-project-member] Lock "6a80e0c1-2088-4de3-9032-d05ff48c3fcb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.960s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.858491] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96c02dc-a02a-4eae-8133-2c7f869919e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.866494] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459cc847-0b5d-4e6d-9bcc-6b44529f3674 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.901388] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d25246-65e5-4dd3-b3da-2a3203aa44f3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.908749] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325c6fa1-a252-467c-ac2c-f53e0d0ae478 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.923295] env[63418]: DEBUG nova.compute.provider_tree [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.122877] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.179066] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 725.427272] env[63418]: DEBUG nova.scheduler.client.report [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 725.470370] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 725.504705] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.504705] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.504705] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.504948] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.504993] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.505654] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.506280] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.506280] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.506493] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.506603] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.506975] env[63418]: DEBUG nova.virt.hardware [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.507812] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49eaf2d9-7011-4274-80f5-f7f8a638f8b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.518866] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a9716b-216e-4367-868b-604141f4b437 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.706799] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.938168] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.938168] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 725.943039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.441s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.015430] env[63418]: DEBUG nova.compute.manager [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Received event network-changed-746f56d6-c17d-4ebd-abcd-da41b3cc1855 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 726.015772] env[63418]: DEBUG nova.compute.manager [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Refreshing instance network info cache due to event network-changed-746f56d6-c17d-4ebd-abcd-da41b3cc1855. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 726.016031] env[63418]: DEBUG oslo_concurrency.lockutils [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] Acquiring lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.016169] env[63418]: DEBUG oslo_concurrency.lockutils [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] Acquired lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.016373] env[63418]: DEBUG nova.network.neutron [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Refreshing network info cache for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.246508] env[63418]: ERROR nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 726.246508] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.246508] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.246508] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.246508] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.246508] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.246508] env[63418]: ERROR nova.compute.manager raise self.value [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.246508] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.246508] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.246508] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.247221] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.247221] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.247221] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 726.247221] env[63418]: ERROR nova.compute.manager [ 726.247221] env[63418]: Traceback (most recent call last): [ 726.247221] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.247221] env[63418]: listener.cb(fileno) [ 726.247221] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.247221] env[63418]: result = function(*args, **kwargs) [ 726.247221] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.247221] env[63418]: return func(*args, **kwargs) [ 726.247221] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.247221] env[63418]: raise e [ 726.247221] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.247221] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 726.247221] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.247221] env[63418]: created_port_ids = self._update_ports_for_instance( [ 726.247221] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.247221] env[63418]: with excutils.save_and_reraise_exception(): [ 726.247221] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.247221] env[63418]: self.force_reraise() [ 726.247221] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.247221] env[63418]: raise self.value [ 726.247221] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.247221] env[63418]: updated_port = self._update_port( [ 726.247221] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.247221] env[63418]: _ensure_no_port_binding_failure(port) [ 726.247221] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.247221] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.248113] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 726.248113] env[63418]: Removing descriptor: 15 [ 726.248113] env[63418]: ERROR nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Traceback (most recent call last): [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] yield resources [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self.driver.spawn(context, instance, image_meta, [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.248113] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] vm_ref = self.build_virtual_machine(instance, [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] for vif in network_info: [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return self._sync_wrapper(fn, *args, **kwargs) [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self.wait() [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self[:] = self._gt.wait() [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return self._exit_event.wait() [ 726.248492] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] result = hub.switch() [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return self.greenlet.switch() [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] result = function(*args, **kwargs) [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return func(*args, **kwargs) [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] raise e [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] nwinfo = self.network_api.allocate_for_instance( [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.248892] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] created_port_ids = self._update_ports_for_instance( [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] with excutils.save_and_reraise_exception(): [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self.force_reraise() [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] raise self.value [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] updated_port = self._update_port( [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] _ensure_no_port_binding_failure(port) [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.249285] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] raise exception.PortBindingFailed(port_id=port['id']) [ 726.249669] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 726.249669] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] [ 726.249669] env[63418]: INFO nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Terminating instance [ 726.454987] env[63418]: DEBUG nova.compute.utils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.461660] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 726.461660] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.534812] env[63418]: DEBUG nova.policy [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72f335c3456347d9b4b10dc21a60588a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a9fab201a5043a8815e9153d80a4f55', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.538438] env[63418]: DEBUG nova.network.neutron [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.718722] env[63418]: DEBUG nova.network.neutron [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.751749] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.941058] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Successfully created port: 1f437bb7-273f-4f91-b781-5c6b845c74cf {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.951336] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d980fb-be0e-4e43-b6a5-88c5eebe20f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.959545] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b1c718-c2d3-4bf0-82fc-f52e45aed742 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.963396] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 726.998613] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac3faea-67a7-46d5-8eaa-ae76d31ff3fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.006853] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a89991d-2377-4e5b-97db-44a59ec74e13 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.020501] env[63418]: DEBUG nova.compute.provider_tree [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.223473] env[63418]: DEBUG oslo_concurrency.lockutils [req-f5a8e489-793e-4898-b23f-a3a52309a8f1 req-26a2f455-2d4b-4b24-b141-7b40849c830c service nova] Releasing lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.223921] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquired lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.224122] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.524686] env[63418]: DEBUG nova.scheduler.client.report [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 727.771100] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.908999] env[63418]: ERROR nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 727.908999] env[63418]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.908999] env[63418]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.908999] env[63418]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.908999] env[63418]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.908999] env[63418]: ERROR nova.compute.manager self.force_reraise() [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.908999] env[63418]: ERROR nova.compute.manager raise self.value [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.908999] env[63418]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.908999] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.908999] env[63418]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.909711] env[63418]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.909711] env[63418]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.909711] env[63418]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 727.909711] env[63418]: ERROR nova.compute.manager [ 727.909711] env[63418]: Traceback (most recent call last): [ 727.909711] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.909711] env[63418]: listener.cb(fileno) [ 727.909711] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.909711] env[63418]: result = function(*args, **kwargs) [ 727.909711] env[63418]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.909711] env[63418]: return func(*args, **kwargs) [ 727.909711] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 727.909711] env[63418]: raise e [ 727.909711] env[63418]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.909711] env[63418]: nwinfo = self.network_api.allocate_for_instance( [ 727.909711] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.909711] env[63418]: created_port_ids = self._update_ports_for_instance( [ 727.909711] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.909711] env[63418]: with excutils.save_and_reraise_exception(): [ 727.909711] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.909711] env[63418]: self.force_reraise() [ 727.909711] env[63418]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.909711] env[63418]: raise self.value [ 727.909711] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.909711] env[63418]: updated_port = self._update_port( [ 727.909711] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.909711] env[63418]: _ensure_no_port_binding_failure(port) [ 727.909711] env[63418]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.909711] env[63418]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.910594] env[63418]: nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 727.910594] env[63418]: Removing descriptor: 15 [ 727.966630] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.004519] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 728.034301] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.034560] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.034734] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.034916] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.039080] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.039306] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.039599] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.039756] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.039967] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.040183] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.040384] env[63418]: DEBUG nova.virt.hardware [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.041795] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.100s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.041795] env[63418]: ERROR nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Traceback (most recent call last): [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self.driver.spawn(context, instance, image_meta, [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.041795] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] vm_ref = self.build_virtual_machine(instance, [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] for vif in network_info: [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return self._sync_wrapper(fn, *args, **kwargs) [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self.wait() [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self[:] = self._gt.wait() [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return self._exit_event.wait() [ 728.042092] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] result = hub.switch() [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return self.greenlet.switch() [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] result = function(*args, **kwargs) [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] return func(*args, **kwargs) [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] raise e [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] nwinfo = self.network_api.allocate_for_instance( [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.042481] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] created_port_ids = self._update_ports_for_instance( [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] with excutils.save_and_reraise_exception(): [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] self.force_reraise() [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] raise self.value [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] updated_port = self._update_port( [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] _ensure_no_port_binding_failure(port) [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.042923] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] raise exception.PortBindingFailed(port_id=port['id']) [ 728.043286] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] nova.exception.PortBindingFailed: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. [ 728.043286] env[63418]: ERROR nova.compute.manager [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] [ 728.043286] env[63418]: DEBUG nova.compute.utils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.045516] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Build of instance c79aa3d1-bbca-45cf-9f29-40a2720c4437 was re-scheduled: Binding failed for port 4eb0a23f-f609-4da1-bf11-a41617596ab5, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 728.045971] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 728.046212] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquiring lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.046360] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Acquired lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.046517] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.048068] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c15eb56-c2ac-4bf2-a30a-3dc27b95d16e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.053239] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.457s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.063236] env[63418]: DEBUG nova.compute.manager [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Received event network-vif-deleted-746f56d6-c17d-4ebd-abcd-da41b3cc1855 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 728.063236] env[63418]: DEBUG nova.compute.manager [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Received event network-changed-1f437bb7-273f-4f91-b781-5c6b845c74cf {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 728.063236] env[63418]: DEBUG nova.compute.manager [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Refreshing instance network info cache due to event network-changed-1f437bb7-273f-4f91-b781-5c6b845c74cf. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 728.063600] env[63418]: DEBUG oslo_concurrency.lockutils [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] Acquiring lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.067107] env[63418]: DEBUG oslo_concurrency.lockutils [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] Acquired lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.067107] env[63418]: DEBUG nova.network.neutron [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Refreshing network info cache for port 1f437bb7-273f-4f91-b781-5c6b845c74cf {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.071061] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eae133c-3bab-4548-b668-fb4ac7a5febe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.091043] env[63418]: ERROR nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] Traceback (most recent call last): [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] yield resources [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self.driver.spawn(context, instance, image_meta, [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] vm_ref = self.build_virtual_machine(instance, [ 728.091043] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] for vif in network_info: [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] return self._sync_wrapper(fn, *args, **kwargs) [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self.wait() [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self[:] = self._gt.wait() [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] return self._exit_event.wait() [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.091478] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] current.throw(*self._exc) [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] result = function(*args, **kwargs) [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] return func(*args, **kwargs) [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] raise e [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] nwinfo = self.network_api.allocate_for_instance( [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] created_port_ids = self._update_ports_for_instance( [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] with excutils.save_and_reraise_exception(): [ 728.092426] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self.force_reraise() [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] raise self.value [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] updated_port = self._update_port( [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] _ensure_no_port_binding_failure(port) [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] raise exception.PortBindingFailed(port_id=port['id']) [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 728.092951] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] [ 728.092951] env[63418]: INFO nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Terminating instance [ 728.472892] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Releasing lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.473633] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 728.473633] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.473886] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-359f6999-d666-4852-b67b-c6b436eaba94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.484041] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f156b4-9de6-4e52-9fa3-a4517c1d3dbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.511598] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7a4d56b-f785-44b4-acb3-6390a4bad83c could not be found. [ 728.511598] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.511598] env[63418]: INFO nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 728.511598] env[63418]: DEBUG oslo.service.loopingcall [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.511598] env[63418]: DEBUG nova.compute.manager [-] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 728.511598] env[63418]: DEBUG nova.network.neutron [-] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.544443] env[63418]: DEBUG nova.network.neutron [-] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.574895] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.598011] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquiring lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.604866] env[63418]: DEBUG nova.network.neutron [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.762150] env[63418]: DEBUG nova.network.neutron [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.836104] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.979943] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31425ed-fd6c-413e-98f3-b59ce59931b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.989513] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc78fe2-7520-4017-9b77-5d31b152c6f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.026019] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1750d8d5-347c-4617-9601-6cfd7627961c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.032086] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954eceb9-593c-49c7-b96a-ed742c9fd110 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.045835] env[63418]: DEBUG nova.compute.provider_tree [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.047435] env[63418]: DEBUG nova.network.neutron [-] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.265833] env[63418]: DEBUG oslo_concurrency.lockutils [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] Releasing lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.266283] env[63418]: DEBUG nova.compute.manager [req-2a0700f0-0fa9-45b6-be48-9a8a0ef30d7d req-15830cc8-b67d-428a-a2bc-95c1796dd63a service nova] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Received event network-vif-deleted-1f437bb7-273f-4f91-b781-5c6b845c74cf {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 729.266837] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquired lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.267247] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.339267] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Releasing lock "refresh_cache-c79aa3d1-bbca-45cf-9f29-40a2720c4437" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.342612] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 729.342612] env[63418]: DEBUG nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 729.342612] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.361968] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.550755] env[63418]: INFO nova.compute.manager [-] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Took 1.04 seconds to deallocate network for instance. [ 729.551697] env[63418]: DEBUG nova.scheduler.client.report [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 729.560242] env[63418]: DEBUG nova.compute.claims [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.560242] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.796754] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.865520] env[63418]: DEBUG nova.network.neutron [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.979463] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.057573] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.005s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.058520] env[63418]: ERROR nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] Traceback (most recent call last): [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self.driver.spawn(context, instance, image_meta, [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] vm_ref = self.build_virtual_machine(instance, [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.058520] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] for vif in network_info: [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return self._sync_wrapper(fn, *args, **kwargs) [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self.wait() [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self[:] = self._gt.wait() [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return self._exit_event.wait() [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] result = hub.switch() [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.058972] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return self.greenlet.switch() [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] result = function(*args, **kwargs) [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] return func(*args, **kwargs) [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] raise e [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] nwinfo = self.network_api.allocate_for_instance( [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] created_port_ids = self._update_ports_for_instance( [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] with excutils.save_and_reraise_exception(): [ 730.059509] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] self.force_reraise() [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] raise self.value [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] updated_port = self._update_port( [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] _ensure_no_port_binding_failure(port) [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] raise exception.PortBindingFailed(port_id=port['id']) [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] nova.exception.PortBindingFailed: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. [ 730.059901] env[63418]: ERROR nova.compute.manager [instance: 96183939-654d-4510-adf2-303f41f94640] [ 730.060289] env[63418]: DEBUG nova.compute.utils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.061624] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.756s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.066215] env[63418]: INFO nova.compute.claims [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.067141] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Build of instance 96183939-654d-4510-adf2-303f41f94640 was re-scheduled: Binding failed for port 50ca70cc-d061-4a2b-9244-01d875347dfd, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 730.067593] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 730.067813] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.067954] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquired lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.069264] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.368266] env[63418]: INFO nova.compute.manager [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] [instance: c79aa3d1-bbca-45cf-9f29-40a2720c4437] Took 1.03 seconds to deallocate network for instance. [ 730.482496] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Releasing lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.482830] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 730.482830] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.483470] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0cc1395-6744-4697-b4b9-b9a3de8de470 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.492883] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235ff36d-16f7-44a1-9f93-97eb33173c25 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.514384] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 702be472-a4bc-4089-94de-4355ee00684a could not be found. [ 730.514687] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.514812] env[63418]: INFO nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 730.515926] env[63418]: DEBUG oslo.service.loopingcall [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.515926] env[63418]: DEBUG nova.compute.manager [-] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 730.515926] env[63418]: DEBUG nova.network.neutron [-] [instance: 702be472-a4bc-4089-94de-4355ee00684a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.536456] env[63418]: DEBUG nova.network.neutron [-] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.591384] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.716448] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.040955] env[63418]: DEBUG nova.network.neutron [-] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.221983] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Releasing lock "refresh_cache-96183939-654d-4510-adf2-303f41f94640" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.222207] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 731.222600] env[63418]: DEBUG nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 731.222600] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.238700] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.411380] env[63418]: INFO nova.scheduler.client.report [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Deleted allocations for instance c79aa3d1-bbca-45cf-9f29-40a2720c4437 [ 731.505808] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4c3a69-b15c-4d8d-b05c-965b7c74afca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.514531] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74107d0-3c8f-4c7d-8d39-6c03d1a1f7aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.546558] env[63418]: INFO nova.compute.manager [-] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Took 1.03 seconds to deallocate network for instance. [ 731.549189] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499bccd8-6c56-4d03-8f71-1933ccb7b264 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.551989] env[63418]: DEBUG nova.compute.claims [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Aborting claim: {{(pid=63418) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.552217] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.557978] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbc907e-b8ba-4e94-8835-6a81b490dc28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.572333] env[63418]: DEBUG nova.compute.provider_tree [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.744439] env[63418]: DEBUG nova.network.neutron [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.920246] env[63418]: DEBUG oslo_concurrency.lockutils [None req-65369f0c-7b7e-49f4-b0d1-427a8c4c739b tempest-ServerActionsTestOtherA-1179747971 tempest-ServerActionsTestOtherA-1179747971-project-member] Lock "c79aa3d1-bbca-45cf-9f29-40a2720c4437" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.203s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.975802] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "f7554ae4-c7a7-4111-a830-10f9029dc074" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.976092] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.075214] env[63418]: DEBUG nova.scheduler.client.report [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 732.248748] env[63418]: INFO nova.compute.manager [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: 96183939-654d-4510-adf2-303f41f94640] Took 1.02 seconds to deallocate network for instance. [ 732.422668] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 732.579747] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.580317] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 732.583353] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.162s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.586577] env[63418]: INFO nova.compute.claims [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.947896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.090633] env[63418]: DEBUG nova.compute.utils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.094517] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 733.094755] env[63418]: DEBUG nova.network.neutron [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.146069] env[63418]: DEBUG nova.policy [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8cab627732b4915b62d1809bd1ad949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4114cfb688a4108adcde1e1bd083ecc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.279579] env[63418]: INFO nova.scheduler.client.report [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Deleted allocations for instance 96183939-654d-4510-adf2-303f41f94640 [ 733.539183] env[63418]: DEBUG nova.network.neutron [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Successfully created port: 9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.599802] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 733.791221] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e8064094-cba0-4124-9a25-9eca8df2b02b tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "96183939-654d-4510-adf2-303f41f94640" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.901s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.988317] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62f2d34-b77e-4075-8c19-886f1ff10b58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.995939] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43841f1-b4a9-4d3e-9fea-25745a8267cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.027436] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbaeb1cc-b26b-4a0e-9117-249b1d8dfff7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.038206] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c2a256-4f40-45a7-a203-a45e1a661b63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.051226] env[63418]: DEBUG nova.compute.provider_tree [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.291768] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 734.554833] env[63418]: DEBUG nova.scheduler.client.report [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 734.608638] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 734.634944] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.635227] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.635381] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.635561] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.635708] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.635853] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.636071] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.636236] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.636409] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.636557] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.636726] env[63418]: DEBUG nova.virt.hardware [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.637894] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c2044e-91f3-46b9-af24-b839d8c01bc0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.646051] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6828d7-e3d6-4293-8553-c60a13ccc884 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.813860] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.060253] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.060798] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 735.063434] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.059s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.063655] env[63418]: DEBUG nova.objects.instance [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 735.171602] env[63418]: DEBUG nova.compute.manager [req-490f3f59-015e-44df-b68b-4e7f2e98d913 req-2dfefaef-e5dd-4fd9-a2dd-7ce72d03c8f6 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Received event network-vif-plugged-9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 735.171857] env[63418]: DEBUG oslo_concurrency.lockutils [req-490f3f59-015e-44df-b68b-4e7f2e98d913 req-2dfefaef-e5dd-4fd9-a2dd-7ce72d03c8f6 service nova] Acquiring lock "6b81dbe2-aa9e-4561-962b-2af167234b90-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.172132] env[63418]: DEBUG oslo_concurrency.lockutils [req-490f3f59-015e-44df-b68b-4e7f2e98d913 req-2dfefaef-e5dd-4fd9-a2dd-7ce72d03c8f6 service nova] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.172321] env[63418]: DEBUG oslo_concurrency.lockutils [req-490f3f59-015e-44df-b68b-4e7f2e98d913 req-2dfefaef-e5dd-4fd9-a2dd-7ce72d03c8f6 service nova] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.172514] env[63418]: DEBUG nova.compute.manager [req-490f3f59-015e-44df-b68b-4e7f2e98d913 req-2dfefaef-e5dd-4fd9-a2dd-7ce72d03c8f6 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] No waiting events found dispatching network-vif-plugged-9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 735.172636] env[63418]: WARNING nova.compute.manager [req-490f3f59-015e-44df-b68b-4e7f2e98d913 req-2dfefaef-e5dd-4fd9-a2dd-7ce72d03c8f6 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Received unexpected event network-vif-plugged-9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 for instance with vm_state building and task_state spawning. [ 735.275822] env[63418]: DEBUG nova.network.neutron [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Successfully updated port: 9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 735.567827] env[63418]: DEBUG nova.compute.utils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.572290] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 735.572475] env[63418]: DEBUG nova.network.neutron [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.608299] env[63418]: DEBUG nova.policy [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8cab627732b4915b62d1809bd1ad949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4114cfb688a4108adcde1e1bd083ecc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.778467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.778467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.778467] env[63418]: DEBUG nova.network.neutron [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.960355] env[63418]: DEBUG nova.network.neutron [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Successfully created port: e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.073147] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 736.076613] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2917fcad-8721-4af5-9585-33a8e45958a7 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.077673] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.708s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.309967] env[63418]: DEBUG nova.network.neutron [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.480649] env[63418]: DEBUG nova.network.neutron [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updating instance_info_cache with network_info: [{"id": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "address": "fa:16:3e:a2:24:98", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d15beed-d1", "ovs_interfaceid": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.895198] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c015766-1904-420c-bd2a-7ec4fb54228b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.902709] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a649e370-a7b0-4356-9d6b-a191d1e70324 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.933270] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d9108c-d004-4cdf-97ce-6c5e27bb317e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.941962] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f946b785-8b7d-43bc-a7f6-6b48dd8f6e63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.955318] env[63418]: DEBUG nova.compute.provider_tree [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.983673] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.983977] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Instance network_info: |[{"id": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "address": "fa:16:3e:a2:24:98", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d15beed-d1", "ovs_interfaceid": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 736.984466] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:24:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d15beed-d1a9-4234-b1ab-fd8df3c8abc6', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.992302] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Creating folder: Project (f4114cfb688a4108adcde1e1bd083ecc). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.992549] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4f15978-1966-4661-892e-a4f9d241b34c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.003256] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Created folder: Project (f4114cfb688a4108adcde1e1bd083ecc) in parent group-v268354. [ 737.003434] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Creating folder: Instances. Parent ref: group-v268387. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.003672] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb71dd9d-f3a2-4ab2-99a2-637018833af4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.012449] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Created folder: Instances in parent group-v268387. [ 737.012716] env[63418]: DEBUG oslo.service.loopingcall [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.012841] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 737.013040] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18bc083a-13fa-4be8-9f9e-ffe35e898e98 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.031883] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.031883] env[63418]: value = "task-1244747" [ 737.031883] env[63418]: _type = "Task" [ 737.031883] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.039053] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244747, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.087832] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 737.115467] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.115713] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.115867] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.116062] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.116207] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.116347] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.116548] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.116705] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.116867] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.117484] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.117761] env[63418]: DEBUG nova.virt.hardware [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.118607] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0119407d-9ec0-452e-9614-76c3cb73c534 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.126809] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8ccfe2-fa88-45f1-9855-ca06e36e54eb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.198201] env[63418]: DEBUG nova.compute.manager [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Received event network-changed-9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 737.198391] env[63418]: DEBUG nova.compute.manager [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Refreshing instance network info cache due to event network-changed-9d15beed-d1a9-4234-b1ab-fd8df3c8abc6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 737.198586] env[63418]: DEBUG oslo_concurrency.lockutils [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] Acquiring lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.198686] env[63418]: DEBUG oslo_concurrency.lockutils [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] Acquired lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.198844] env[63418]: DEBUG nova.network.neutron [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Refreshing network info cache for port 9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 737.458216] env[63418]: DEBUG nova.scheduler.client.report [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 737.513358] env[63418]: DEBUG nova.network.neutron [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Successfully updated port: e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 737.541853] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244747, 'name': CreateVM_Task, 'duration_secs': 0.30289} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.542027] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 737.549192] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.549406] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.549739] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.549988] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f2b3813-b0ec-4d0b-b7fc-2a4622f6d887 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.554664] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 737.554664] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c05ffb-38b6-bf1e-95ab-c218d9087913" [ 737.554664] env[63418]: _type = "Task" [ 737.554664] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.562770] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c05ffb-38b6-bf1e-95ab-c218d9087913, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.963155] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.885s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.963833] env[63418]: ERROR nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Traceback (most recent call last): [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self.driver.spawn(context, instance, image_meta, [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] vm_ref = self.build_virtual_machine(instance, [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.963833] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] for vif in network_info: [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return self._sync_wrapper(fn, *args, **kwargs) [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self.wait() [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self[:] = self._gt.wait() [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return self._exit_event.wait() [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] result = hub.switch() [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.965180] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return self.greenlet.switch() [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] result = function(*args, **kwargs) [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] return func(*args, **kwargs) [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] raise e [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] nwinfo = self.network_api.allocate_for_instance( [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] created_port_ids = self._update_ports_for_instance( [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] with excutils.save_and_reraise_exception(): [ 737.965560] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] self.force_reraise() [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] raise self.value [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] updated_port = self._update_port( [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] _ensure_no_port_binding_failure(port) [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] raise exception.PortBindingFailed(port_id=port['id']) [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] nova.exception.PortBindingFailed: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. [ 737.965981] env[63418]: ERROR nova.compute.manager [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] [ 737.966319] env[63418]: DEBUG nova.compute.utils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.966319] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.156s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.967450] env[63418]: INFO nova.compute.claims [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.970121] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Build of instance 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3 was re-scheduled: Binding failed for port 47e3f4ac-5b56-4910-9b8a-c942fa3a7cd2, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 737.970555] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 737.970774] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquiring lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.970919] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Acquired lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.971088] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.015477] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "refresh_cache-17a055e4-13da-4984-ae64-af1ae9d652f6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.015632] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "refresh_cache-17a055e4-13da-4984-ae64-af1ae9d652f6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.015799] env[63418]: DEBUG nova.network.neutron [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.065531] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c05ffb-38b6-bf1e-95ab-c218d9087913, 'name': SearchDatastore_Task, 'duration_secs': 0.008809} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.068380] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.068747] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.069146] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.069369] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.069699] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.069967] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5413a5e-0fd1-45f5-b4ff-6249ca61169e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.078626] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.078626] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 738.079253] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ce7d0e8-13c2-4a3a-ba23-973261303c76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.084210] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 738.084210] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5252a3a0-ee24-ce5b-a507-92bb1382dc74" [ 738.084210] env[63418]: _type = "Task" [ 738.084210] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.092342] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5252a3a0-ee24-ce5b-a507-92bb1382dc74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.208028] env[63418]: DEBUG nova.network.neutron [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updated VIF entry in instance network info cache for port 9d15beed-d1a9-4234-b1ab-fd8df3c8abc6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 738.208975] env[63418]: DEBUG nova.network.neutron [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updating instance_info_cache with network_info: [{"id": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "address": "fa:16:3e:a2:24:98", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d15beed-d1", "ovs_interfaceid": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.490998] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.548333] env[63418]: DEBUG nova.network.neutron [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.591946] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.601219] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5252a3a0-ee24-ce5b-a507-92bb1382dc74, 'name': SearchDatastore_Task, 'duration_secs': 0.00853} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.605024] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d04b3082-06fb-4bb4-862f-d634a82e922e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.608456] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 738.608456] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5274bf57-5409-f7e6-419f-5327b2d2793a" [ 738.608456] env[63418]: _type = "Task" [ 738.608456] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.616165] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5274bf57-5409-f7e6-419f-5327b2d2793a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.684188] env[63418]: DEBUG nova.network.neutron [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Updating instance_info_cache with network_info: [{"id": "e834b79c-1139-41ba-92e2-9875b4e19a5b", "address": "fa:16:3e:9b:8f:8d", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape834b79c-11", "ovs_interfaceid": "e834b79c-1139-41ba-92e2-9875b4e19a5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.714931] env[63418]: DEBUG oslo_concurrency.lockutils [req-889ffe14-4a98-4b95-bf89-10595c427ff4 req-869458c7-3b23-4b06-a3b7-9d49dd3701e8 service nova] Releasing lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.095083] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Releasing lock "refresh_cache-2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.095367] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 739.095720] env[63418]: DEBUG nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 739.095720] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.115661] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.125494] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5274bf57-5409-f7e6-419f-5327b2d2793a, 'name': SearchDatastore_Task, 'duration_secs': 0.009436} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.126366] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.126626] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 6b81dbe2-aa9e-4561-962b-2af167234b90/6b81dbe2-aa9e-4561-962b-2af167234b90.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.126894] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93d0c6a9-2f23-49a9-8e00-460267321f22 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.136696] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 739.136696] env[63418]: value = "task-1244748" [ 739.136696] env[63418]: _type = "Task" [ 739.136696] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.147197] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.189485] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "refresh_cache-17a055e4-13da-4984-ae64-af1ae9d652f6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.189785] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Instance network_info: |[{"id": "e834b79c-1139-41ba-92e2-9875b4e19a5b", "address": "fa:16:3e:9b:8f:8d", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape834b79c-11", "ovs_interfaceid": "e834b79c-1139-41ba-92e2-9875b4e19a5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 739.190205] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:8f:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e834b79c-1139-41ba-92e2-9875b4e19a5b', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.197694] env[63418]: DEBUG oslo.service.loopingcall [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.200721] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 739.201826] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73b72ca0-128b-4d40-84d9-b38c13db4a77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.223638] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.223638] env[63418]: value = "task-1244749" [ 739.223638] env[63418]: _type = "Task" [ 739.223638] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.231120] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244749, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.236464] env[63418]: DEBUG nova.compute.manager [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Received event network-vif-plugged-e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 739.236464] env[63418]: DEBUG oslo_concurrency.lockutils [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] Acquiring lock "17a055e4-13da-4984-ae64-af1ae9d652f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.236464] env[63418]: DEBUG oslo_concurrency.lockutils [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.236464] env[63418]: DEBUG oslo_concurrency.lockutils [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.236464] env[63418]: DEBUG nova.compute.manager [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] No waiting events found dispatching network-vif-plugged-e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 739.236654] env[63418]: WARNING nova.compute.manager [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Received unexpected event network-vif-plugged-e834b79c-1139-41ba-92e2-9875b4e19a5b for instance with vm_state building and task_state spawning. [ 739.236654] env[63418]: DEBUG nova.compute.manager [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Received event network-changed-e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 739.236725] env[63418]: DEBUG nova.compute.manager [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Refreshing instance network info cache due to event network-changed-e834b79c-1139-41ba-92e2-9875b4e19a5b. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 739.236847] env[63418]: DEBUG oslo_concurrency.lockutils [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] Acquiring lock "refresh_cache-17a055e4-13da-4984-ae64-af1ae9d652f6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.236980] env[63418]: DEBUG oslo_concurrency.lockutils [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] Acquired lock "refresh_cache-17a055e4-13da-4984-ae64-af1ae9d652f6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.237149] env[63418]: DEBUG nova.network.neutron [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Refreshing network info cache for port e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.320891] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edc6e3b-d4ad-499a-ba1c-1c34d93448b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.328906] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991fc064-289a-4a12-a111-b44846b77b51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.359509] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41722ddb-7a40-46d4-b5ab-7389d5697fdf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.367503] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2736c468-3e96-43b2-84ba-6d8fedd9dfd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.383833] env[63418]: DEBUG nova.compute.provider_tree [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.619913] env[63418]: DEBUG nova.network.neutron [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.645920] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491728} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.646230] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 6b81dbe2-aa9e-4561-962b-2af167234b90/6b81dbe2-aa9e-4561-962b-2af167234b90.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 739.646386] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 739.646650] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0ccc3db-6462-47f1-8aa9-a9ed44e9e1cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.652142] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 739.652142] env[63418]: value = "task-1244750" [ 739.652142] env[63418]: _type = "Task" [ 739.652142] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.659820] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.732886] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244749, 'name': CreateVM_Task, 'duration_secs': 0.38162} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.733077] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 739.733760] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.733933] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.734270] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 739.734508] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0e7a976-d588-4683-b8fe-cb2296cb550d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.738786] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 739.738786] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc5b0c-7b05-f9db-1bb3-d05bec293fb5" [ 739.738786] env[63418]: _type = "Task" [ 739.738786] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.747856] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc5b0c-7b05-f9db-1bb3-d05bec293fb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.887294] env[63418]: DEBUG nova.scheduler.client.report [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 739.936351] env[63418]: DEBUG nova.network.neutron [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Updated VIF entry in instance network info cache for port e834b79c-1139-41ba-92e2-9875b4e19a5b. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 739.936708] env[63418]: DEBUG nova.network.neutron [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Updating instance_info_cache with network_info: [{"id": "e834b79c-1139-41ba-92e2-9875b4e19a5b", "address": "fa:16:3e:9b:8f:8d", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape834b79c-11", "ovs_interfaceid": "e834b79c-1139-41ba-92e2-9875b4e19a5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.122555] env[63418]: INFO nova.compute.manager [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] [instance: 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3] Took 1.03 seconds to deallocate network for instance. [ 740.162209] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067202} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.163145] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.163930] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3c09a6-ab03-4a6b-8971-4aa2ef2c079f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.186248] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 6b81dbe2-aa9e-4561-962b-2af167234b90/6b81dbe2-aa9e-4561-962b-2af167234b90.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.186767] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aafec266-494d-49f9-9f68-6b2f40fff682 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.206337] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 740.206337] env[63418]: value = "task-1244751" [ 740.206337] env[63418]: _type = "Task" [ 740.206337] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.215164] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244751, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.248798] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc5b0c-7b05-f9db-1bb3-d05bec293fb5, 'name': SearchDatastore_Task, 'duration_secs': 0.030412} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.249174] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.250418] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.250679] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.250830] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.251023] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.251674] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1a7d101-f3ee-4cc4-ba00-ed4f259ea64c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.259277] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.259445] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 740.260138] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67437808-f378-40b2-b1a3-1e2e8ec20209 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.266182] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 740.266182] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5258f4ee-8a52-f8f5-83b4-aa1f18210e2d" [ 740.266182] env[63418]: _type = "Task" [ 740.266182] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.275231] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5258f4ee-8a52-f8f5-83b4-aa1f18210e2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.391917] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.392499] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 740.395109] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.800s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.395330] env[63418]: DEBUG nova.objects.instance [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lazy-loading 'resources' on Instance uuid 99d00df2-6bf3-4ffe-b77d-f44ada631c8b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.439021] env[63418]: DEBUG oslo_concurrency.lockutils [req-0c889aa7-d4c9-4140-a4b0-695127941b21 req-26899f7e-e52c-4e62-bdee-393b173dfea2 service nova] Releasing lock "refresh_cache-17a055e4-13da-4984-ae64-af1ae9d652f6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.716533] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244751, 'name': ReconfigVM_Task, 'duration_secs': 0.287082} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.716984] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 6b81dbe2-aa9e-4561-962b-2af167234b90/6b81dbe2-aa9e-4561-962b-2af167234b90.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.717766] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f475eed6-8841-49c9-9480-aef1cc9474ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.723658] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 740.723658] env[63418]: value = "task-1244752" [ 740.723658] env[63418]: _type = "Task" [ 740.723658] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.731673] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244752, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.775372] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5258f4ee-8a52-f8f5-83b4-aa1f18210e2d, 'name': SearchDatastore_Task, 'duration_secs': 0.009528} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.776204] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c39fc71-518d-4ced-a17e-79faeeefd5f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.781256] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 740.781256] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea606-0778-5aeb-76df-7ded542b9a78" [ 740.781256] env[63418]: _type = "Task" [ 740.781256] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.788439] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea606-0778-5aeb-76df-7ded542b9a78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.898545] env[63418]: DEBUG nova.compute.utils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.902575] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 740.902739] env[63418]: DEBUG nova.network.neutron [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.940445] env[63418]: DEBUG nova.policy [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8cab627732b4915b62d1809bd1ad949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4114cfb688a4108adcde1e1bd083ecc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.168379] env[63418]: INFO nova.scheduler.client.report [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Deleted allocations for instance 2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3 [ 741.195101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35281575-7fe7-4469-87e4-a5a5912d22e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.207756] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a52a345-08e1-45cc-a9f8-72a2019efcae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.210428] env[63418]: DEBUG nova.network.neutron [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Successfully created port: 1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.244709] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb11ac09-d7ae-42db-946f-1b54986f04d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.256056] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244752, 'name': Rename_Task, 'duration_secs': 0.135153} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.260594] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4f6c7f-c93b-489b-8816-fc58e10d0bb7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.262876] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 741.262876] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94bf3f06-26ee-427f-a1c9-66b407362c59 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.273264] env[63418]: DEBUG nova.compute.provider_tree [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.275495] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 741.275495] env[63418]: value = "task-1244753" [ 741.275495] env[63418]: _type = "Task" [ 741.275495] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.287025] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244753, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.293169] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea606-0778-5aeb-76df-7ded542b9a78, 'name': SearchDatastore_Task, 'duration_secs': 0.009209} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.293419] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.293716] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 17a055e4-13da-4984-ae64-af1ae9d652f6/17a055e4-13da-4984-ae64-af1ae9d652f6.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 741.293966] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d373b4bc-a9f2-41e2-b159-04e070c4d18d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.300662] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 741.300662] env[63418]: value = "task-1244754" [ 741.300662] env[63418]: _type = "Task" [ 741.300662] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.309656] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244754, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.404020] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 741.679537] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e36dd502-6e4e-4469-9c68-6ebc20587bb0 tempest-ServersTestFqdnHostnames-2092181538 tempest-ServersTestFqdnHostnames-2092181538-project-member] Lock "2dc02ba1-e8d9-42de-9a99-dbdbf53e39e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.784s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.777154] env[63418]: DEBUG nova.scheduler.client.report [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 741.789901] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244753, 'name': PowerOnVM_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.810553] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244754, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477206} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.810553] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 17a055e4-13da-4984-ae64-af1ae9d652f6/17a055e4-13da-4984-ae64-af1ae9d652f6.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.810797] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.810968] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbb257ac-293c-4667-8703-e883f7a86825 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.816687] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 741.816687] env[63418]: value = "task-1244755" [ 741.816687] env[63418]: _type = "Task" [ 741.816687] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.823992] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244755, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.182214] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 742.285482] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.287567] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.560s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.289152] env[63418]: INFO nova.compute.claims [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.299399] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244753, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.309906] env[63418]: INFO nova.scheduler.client.report [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Deleted allocations for instance 99d00df2-6bf3-4ffe-b77d-f44ada631c8b [ 742.326465] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244755, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.26072} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.326731] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.327599] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5222f7-af34-4332-a805-ef7ec14c0841 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.351381] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 17a055e4-13da-4984-ae64-af1ae9d652f6/17a055e4-13da-4984-ae64-af1ae9d652f6.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.353031] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7355e901-1d61-4f2f-9d4c-6ec5de72174e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.373074] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 742.373074] env[63418]: value = "task-1244756" [ 742.373074] env[63418]: _type = "Task" [ 742.373074] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.381697] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244756, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.413078] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 742.441744] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.441959] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.442132] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.442314] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.442457] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.442602] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.443016] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.443016] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.444154] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.444154] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.444154] env[63418]: DEBUG nova.virt.hardware [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.444391] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eefdd7d-1216-47ad-aa60-297ecd56b1b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.453037] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c0ced0-67a4-4a5f-aec9-092fca640ad7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.705231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.794019] env[63418]: DEBUG oslo_vmware.api [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244753, 'name': PowerOnVM_Task, 'duration_secs': 1.08239} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.794019] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 742.794019] env[63418]: INFO nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Took 8.18 seconds to spawn the instance on the hypervisor. [ 742.794019] env[63418]: DEBUG nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 742.794019] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a031d0-8e26-47f0-9a25-8a83cd3b6076 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.816974] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83702b4f-c904-48a6-b00e-9a11cc9eada9 tempest-ServerShowV254Test-231049312 tempest-ServerShowV254Test-231049312-project-member] Lock "99d00df2-6bf3-4ffe-b77d-f44ada631c8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.466s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.883247] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244756, 'name': ReconfigVM_Task, 'duration_secs': 0.30956} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.885023] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 17a055e4-13da-4984-ae64-af1ae9d652f6/17a055e4-13da-4984-ae64-af1ae9d652f6.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.885023] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2cea48f-c837-41d9-9748-587385c5000e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.894217] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 742.894217] env[63418]: value = "task-1244757" [ 742.894217] env[63418]: _type = "Task" [ 742.894217] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.900254] env[63418]: DEBUG nova.compute.manager [req-06ef18dc-309a-4c86-ac45-de4a4e29265e req-95cfb7e7-853f-4f46-acb2-6c8ce1b19cd0 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Received event network-vif-plugged-1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 742.900254] env[63418]: DEBUG oslo_concurrency.lockutils [req-06ef18dc-309a-4c86-ac45-de4a4e29265e req-95cfb7e7-853f-4f46-acb2-6c8ce1b19cd0 service nova] Acquiring lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.902070] env[63418]: DEBUG oslo_concurrency.lockutils [req-06ef18dc-309a-4c86-ac45-de4a4e29265e req-95cfb7e7-853f-4f46-acb2-6c8ce1b19cd0 service nova] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.902070] env[63418]: DEBUG oslo_concurrency.lockutils [req-06ef18dc-309a-4c86-ac45-de4a4e29265e req-95cfb7e7-853f-4f46-acb2-6c8ce1b19cd0 service nova] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.902070] env[63418]: DEBUG nova.compute.manager [req-06ef18dc-309a-4c86-ac45-de4a4e29265e req-95cfb7e7-853f-4f46-acb2-6c8ce1b19cd0 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] No waiting events found dispatching network-vif-plugged-1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 742.902070] env[63418]: WARNING nova.compute.manager [req-06ef18dc-309a-4c86-ac45-de4a4e29265e req-95cfb7e7-853f-4f46-acb2-6c8ce1b19cd0 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Received unexpected event network-vif-plugged-1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 for instance with vm_state building and task_state spawning. [ 742.907305] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244757, 'name': Rename_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.988953] env[63418]: DEBUG nova.network.neutron [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Successfully updated port: 1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 743.322325] env[63418]: INFO nova.compute.manager [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Took 28.04 seconds to build instance. [ 743.405043] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244757, 'name': Rename_Task, 'duration_secs': 0.145465} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.412016] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 743.412016] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1adc6c20-af2f-4611-83c6-2cfcd97507f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.421712] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 743.421712] env[63418]: value = "task-1244758" [ 743.421712] env[63418]: _type = "Task" [ 743.421712] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.430838] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.491701] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "refresh_cache-c0fd1999-346d-4c12-9b8c-ab7e21ec4227" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.492091] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "refresh_cache-c0fd1999-346d-4c12-9b8c-ab7e21ec4227" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.492091] env[63418]: DEBUG nova.network.neutron [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.671017] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c7486c-d418-4781-966f-eee37852a377 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.682030] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1319374-3c2c-4cb4-9f55-d9b3dc52da8b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.711128] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07823cd-d263-4d36-8ff2-9160d7224d95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.720224] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355a1c09-f3b4-4a82-894b-c1fc480c222b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.736821] env[63418]: DEBUG nova.compute.provider_tree [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.827447] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b9ddfcee-d821-4542-b9c8-de3444fcb5da tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.914s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.937956] env[63418]: DEBUG oslo_vmware.api [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244758, 'name': PowerOnVM_Task, 'duration_secs': 0.472528} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.938244] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 743.938435] env[63418]: INFO nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Took 6.85 seconds to spawn the instance on the hypervisor. [ 743.938610] env[63418]: DEBUG nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 743.939446] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc121085-f9ec-45a3-9afd-488746deb9e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.031555] env[63418]: DEBUG nova.network.neutron [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.174509] env[63418]: DEBUG nova.network.neutron [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Updating instance_info_cache with network_info: [{"id": "1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99", "address": "fa:16:3e:10:ea:f8", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cf7b4bf-b1", "ovs_interfaceid": "1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.240105] env[63418]: DEBUG nova.scheduler.client.report [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 744.330410] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 744.466229] env[63418]: INFO nova.compute.manager [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Took 27.07 seconds to build instance. [ 744.679035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "refresh_cache-c0fd1999-346d-4c12-9b8c-ab7e21ec4227" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.679035] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Instance network_info: |[{"id": "1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99", "address": "fa:16:3e:10:ea:f8", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cf7b4bf-b1", "ovs_interfaceid": "1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 744.679331] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:ea:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.690537] env[63418]: DEBUG oslo.service.loopingcall [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.690911] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 744.691560] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e87d8103-7b9f-4df3-91db-9c799424d063 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.715531] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.715531] env[63418]: value = "task-1244759" [ 744.715531] env[63418]: _type = "Task" [ 744.715531] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.723921] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244759, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.745904] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.746505] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 744.749367] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.627s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.749549] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.750254] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 744.750254] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.043s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.751396] env[63418]: INFO nova.compute.claims [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.759273] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d593dcc-25d4-4ac3-9263-812daa87a129 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.771448] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019d4ead-6edf-4c5e-94fb-9fee8ff9729a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.789665] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd6f2c4-75e2-4ce8-9292-f0111e13e935 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.794732] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f724217-2714-4f29-b217-3bdce382b202 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.828542] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181534MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 744.828712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.857165] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.969532] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5a563eef-a85d-43a8-9a62-57507d0e0800 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.616s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.982189] env[63418]: DEBUG nova.compute.manager [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Received event network-changed-1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 744.982374] env[63418]: DEBUG nova.compute.manager [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Refreshing instance network info cache due to event network-changed-1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 744.983322] env[63418]: DEBUG oslo_concurrency.lockutils [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] Acquiring lock "refresh_cache-c0fd1999-346d-4c12-9b8c-ab7e21ec4227" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.983322] env[63418]: DEBUG oslo_concurrency.lockutils [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] Acquired lock "refresh_cache-c0fd1999-346d-4c12-9b8c-ab7e21ec4227" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.983322] env[63418]: DEBUG nova.network.neutron [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Refreshing network info cache for port 1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 745.226073] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244759, 'name': CreateVM_Task, 'duration_secs': 0.341914} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.226273] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 745.226948] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.227125] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.227462] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 745.227721] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78e9721d-98b9-45ae-9bcb-55c7d5e65a4b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.232409] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 745.232409] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5229261c-fc0e-61ab-bb3a-96164ab86b09" [ 745.232409] env[63418]: _type = "Task" [ 745.232409] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.240310] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5229261c-fc0e-61ab-bb3a-96164ab86b09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.260638] env[63418]: DEBUG nova.compute.utils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.262044] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 745.262829] env[63418]: DEBUG nova.network.neutron [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.334464] env[63418]: DEBUG nova.policy [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '674a82efba9d4d8ca86e52640f9a5589', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e9f537407b84d50a49600de59e72c86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.473541] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 745.734407] env[63418]: DEBUG nova.network.neutron [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Updated VIF entry in instance network info cache for port 1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 745.735042] env[63418]: DEBUG nova.network.neutron [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Updating instance_info_cache with network_info: [{"id": "1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99", "address": "fa:16:3e:10:ea:f8", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cf7b4bf-b1", "ovs_interfaceid": "1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.748752] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5229261c-fc0e-61ab-bb3a-96164ab86b09, 'name': SearchDatastore_Task, 'duration_secs': 0.01743} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.748752] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.748752] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.748752] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.749847] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.749847] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.749847] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04844091-a91f-4b6d-ab0f-09dcfb8641b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.760435] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.760628] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 745.761389] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e15e9bd1-41f3-4ea2-a5e9-e4b9e20b0115 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.764564] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 745.771709] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 745.771709] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521e7cb3-7f32-3cdd-2513-2fec4de7e935" [ 745.771709] env[63418]: _type = "Task" [ 745.771709] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.783255] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521e7cb3-7f32-3cdd-2513-2fec4de7e935, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.912126] env[63418]: DEBUG nova.network.neutron [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Successfully created port: 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.996131] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.153017] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1354d94c-7e87-46dc-bd72-72f1166d3896 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.160580] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41dd84c-0897-481c-a216-30b6aed33b19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.194030] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31cc534-e5d2-443e-824d-7422d69947d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.200388] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47deaffe-2e58-47fe-a118-138e03c10e9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.219198] env[63418]: DEBUG nova.compute.provider_tree [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.242289] env[63418]: DEBUG oslo_concurrency.lockutils [req-17c47f97-3cbe-4227-ad43-7ba1e3edaeca req-5fb9f510-edc7-4296-a521-7e0bcd3a8ab2 service nova] Releasing lock "refresh_cache-c0fd1999-346d-4c12-9b8c-ab7e21ec4227" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.281964] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521e7cb3-7f32-3cdd-2513-2fec4de7e935, 'name': SearchDatastore_Task, 'duration_secs': 0.017604} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.282942] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e3b08eb-2a82-46bf-b676-cc2ea0dbd0e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.291286] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 746.291286] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5279bc14-3b64-86fd-9bbc-4e72f504104e" [ 746.291286] env[63418]: _type = "Task" [ 746.291286] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.301515] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5279bc14-3b64-86fd-9bbc-4e72f504104e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.722034] env[63418]: DEBUG nova.scheduler.client.report [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 746.773855] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 746.807924] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5279bc14-3b64-86fd-9bbc-4e72f504104e, 'name': SearchDatastore_Task, 'duration_secs': 0.024857} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.810252] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.810518] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.810761] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.811031] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.811212] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.811396] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.811643] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.813152] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.813232] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.813399] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.813629] env[63418]: DEBUG nova.virt.hardware [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.813903] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.814256] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] c0fd1999-346d-4c12-9b8c-ab7e21ec4227/c0fd1999-346d-4c12-9b8c-ab7e21ec4227.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 746.815283] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d29bc38-92e5-4694-b29c-595d6e0b96a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.818249] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-19d02f50-caeb-4c4c-a6d0-b24c182aeb7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.827900] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53705272-c4c4-497c-940f-86f7311fc119 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.832234] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 746.832234] env[63418]: value = "task-1244760" [ 746.832234] env[63418]: _type = "Task" [ 746.832234] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.848461] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.231246] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.231246] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 747.232849] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.674s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.348810] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244760, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.454352] env[63418]: DEBUG nova.compute.manager [req-8496db1a-1046-4f3e-8c1f-25afb116986e req-c8f69851-325d-45f2-bde1-30dd1cb5593b service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-vif-plugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 747.454589] env[63418]: DEBUG oslo_concurrency.lockutils [req-8496db1a-1046-4f3e-8c1f-25afb116986e req-c8f69851-325d-45f2-bde1-30dd1cb5593b service nova] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.454812] env[63418]: DEBUG oslo_concurrency.lockutils [req-8496db1a-1046-4f3e-8c1f-25afb116986e req-c8f69851-325d-45f2-bde1-30dd1cb5593b service nova] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.455071] env[63418]: DEBUG oslo_concurrency.lockutils [req-8496db1a-1046-4f3e-8c1f-25afb116986e req-c8f69851-325d-45f2-bde1-30dd1cb5593b service nova] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.455215] env[63418]: DEBUG nova.compute.manager [req-8496db1a-1046-4f3e-8c1f-25afb116986e req-c8f69851-325d-45f2-bde1-30dd1cb5593b service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] No waiting events found dispatching network-vif-plugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 747.455382] env[63418]: WARNING nova.compute.manager [req-8496db1a-1046-4f3e-8c1f-25afb116986e req-c8f69851-325d-45f2-bde1-30dd1cb5593b service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received unexpected event network-vif-plugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 for instance with vm_state building and task_state spawning. [ 747.558455] env[63418]: DEBUG nova.network.neutron [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Successfully updated port: 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 747.740855] env[63418]: DEBUG nova.compute.utils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.743027] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 747.743225] env[63418]: DEBUG nova.network.neutron [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.789553] env[63418]: DEBUG nova.policy [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ff64612530b451fb41100a5aa601be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d52f1fceb24234a8d967038b43c857', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.844313] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244760, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612341} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.846805] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] c0fd1999-346d-4c12-9b8c-ab7e21ec4227/c0fd1999-346d-4c12-9b8c-ab7e21ec4227.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 747.847036] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.847492] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5126c655-1bb6-4636-9ebb-71772132befd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.855264] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 747.855264] env[63418]: value = "task-1244761" [ 747.855264] env[63418]: _type = "Task" [ 747.855264] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.866018] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244761, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.062685] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.062685] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.062685] env[63418]: DEBUG nova.network.neutron [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.073767] env[63418]: DEBUG nova.network.neutron [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Successfully created port: b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.078557] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1ffe8d-1872-4650-bfc4-3596312576d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.089121] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25729bb4-bb89-4976-8f78-879dcddd31f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.130122] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe1d044-eb24-4b32-9a7e-40880750d218 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.138129] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488e825c-0b7a-45e5-ab29-8e5c9655567f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.152464] env[63418]: DEBUG nova.compute.provider_tree [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.244788] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 748.366171] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244761, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090804} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.366443] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.367226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca9b4ee-c1d9-4801-9764-f68ff5e4a1be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.389167] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] c0fd1999-346d-4c12-9b8c-ab7e21ec4227/c0fd1999-346d-4c12-9b8c-ab7e21ec4227.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.389514] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b45e9ca-b884-41ee-9b92-8e16718addc3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.413020] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 748.413020] env[63418]: value = "task-1244762" [ 748.413020] env[63418]: _type = "Task" [ 748.413020] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.420203] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244762, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.500908] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.501369] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.596287] env[63418]: DEBUG nova.network.neutron [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.656013] env[63418]: DEBUG nova.scheduler.client.report [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 748.816328] env[63418]: DEBUG nova.network.neutron [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.920559] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244762, 'name': ReconfigVM_Task, 'duration_secs': 0.373968} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.920840] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Reconfigured VM instance instance-00000030 to attach disk [datastore1] c0fd1999-346d-4c12-9b8c-ab7e21ec4227/c0fd1999-346d-4c12-9b8c-ab7e21ec4227.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.921476] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a3b6a7e-1e6a-40f8-b30a-2c081aad7b69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.928166] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 748.928166] env[63418]: value = "task-1244763" [ 748.928166] env[63418]: _type = "Task" [ 748.928166] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.935797] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244763, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.161812] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.162501] env[63418]: ERROR nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Traceback (most recent call last): [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self.driver.spawn(context, instance, image_meta, [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] vm_ref = self.build_virtual_machine(instance, [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.162501] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] for vif in network_info: [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return self._sync_wrapper(fn, *args, **kwargs) [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self.wait() [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self[:] = self._gt.wait() [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return self._exit_event.wait() [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] result = hub.switch() [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.163175] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return self.greenlet.switch() [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] result = function(*args, **kwargs) [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] return func(*args, **kwargs) [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] raise e [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] nwinfo = self.network_api.allocate_for_instance( [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] created_port_ids = self._update_ports_for_instance( [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] with excutils.save_and_reraise_exception(): [ 749.163847] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] self.force_reraise() [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] raise self.value [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] updated_port = self._update_port( [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] _ensure_no_port_binding_failure(port) [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] raise exception.PortBindingFailed(port_id=port['id']) [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] nova.exception.PortBindingFailed: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. [ 749.164581] env[63418]: ERROR nova.compute.manager [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] [ 749.165139] env[63418]: DEBUG nova.compute.utils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.165139] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.612s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.167694] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Build of instance e7a4d56b-f785-44b4-acb3-6390a4bad83c was re-scheduled: Binding failed for port 746f56d6-c17d-4ebd-abcd-da41b3cc1855, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 749.172285] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 749.172565] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquiring lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.172717] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Acquired lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.172880] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.263324] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 749.290403] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.290698] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.290856] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.291050] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.291198] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.291343] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.291547] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.291732] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.291912] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.292086] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.292259] env[63418]: DEBUG nova.virt.hardware [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.293174] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6775bf-61ac-4f0f-a524-32765d873b64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.302478] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f05ff9-e819-451b-92a6-7b1a5683a1ad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.321933] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.322702] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance network_info: |[{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 749.323066] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:37:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0263ccbe-8541-4cf4-bd2a-0e9b517d6f29', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.330495] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating folder: Project (0e9f537407b84d50a49600de59e72c86). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 749.330769] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5df6765b-428d-4117-9089-801b891bf670 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.341329] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created folder: Project (0e9f537407b84d50a49600de59e72c86) in parent group-v268354. [ 749.341507] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating folder: Instances. Parent ref: group-v268392. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 749.341772] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e320160d-413c-45df-bb0b-66e3400b03b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.350598] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created folder: Instances in parent group-v268392. [ 749.350833] env[63418]: DEBUG oslo.service.loopingcall [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.351029] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 749.351238] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b0cb3bc-23c9-418c-8561-c109310bde41 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.371917] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.371917] env[63418]: value = "task-1244766" [ 749.371917] env[63418]: _type = "Task" [ 749.371917] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.380101] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244766, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.440629] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244763, 'name': Rename_Task, 'duration_secs': 0.14283} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.441378] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 749.441725] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4348a3c-f6e4-4fb0-b671-3e86efb9a462 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.447982] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 749.447982] env[63418]: value = "task-1244767" [ 749.447982] env[63418]: _type = "Task" [ 749.447982] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.456645] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.500132] env[63418]: DEBUG nova.compute.manager [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 749.500132] env[63418]: DEBUG nova.compute.manager [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing instance network info cache due to event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 749.500132] env[63418]: DEBUG oslo_concurrency.lockutils [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.500132] env[63418]: DEBUG oslo_concurrency.lockutils [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.500132] env[63418]: DEBUG nova.network.neutron [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.597304] env[63418]: DEBUG nova.network.neutron [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Successfully updated port: b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 749.694223] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.772580] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.882218] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244766, 'name': CreateVM_Task, 'duration_secs': 0.354626} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.884692] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 749.886010] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.886010] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.886157] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 749.887046] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dbee100-3953-4d0d-b92a-5b0a7ee2e58b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.891271] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 749.891271] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]524aae15-5d2b-7eab-3ebe-e06ecaf8b126" [ 749.891271] env[63418]: _type = "Task" [ 749.891271] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.905157] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524aae15-5d2b-7eab-3ebe-e06ecaf8b126, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.957914] env[63418]: DEBUG oslo_vmware.api [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244767, 'name': PowerOnVM_Task, 'duration_secs': 0.475458} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.958204] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.958402] env[63418]: INFO nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Took 7.55 seconds to spawn the instance on the hypervisor. [ 749.958577] env[63418]: DEBUG nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 749.959603] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33bd4cd-876d-4192-a84a-e577a82cd531 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.006892] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9bac80-2271-4c41-91f6-bb3dfde68bbd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.015586] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669dfdc1-a34a-4b35-8507-d486274a38a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.051902] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc74d1a-df2e-4f8a-88f4-a84fa7c7e7d5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.060865] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7496981-0251-4a38-8fd2-f056780ff57f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.076761] env[63418]: DEBUG nova.compute.provider_tree [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.099512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-c4cb1770-b3c7-4f6b-ba11-85310f2175a0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.099648] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-c4cb1770-b3c7-4f6b-ba11-85310f2175a0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.099795] env[63418]: DEBUG nova.network.neutron [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.247272] env[63418]: DEBUG nova.network.neutron [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updated VIF entry in instance network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 750.247611] env[63418]: DEBUG nova.network.neutron [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.274867] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Releasing lock "refresh_cache-e7a4d56b-f785-44b4-acb3-6390a4bad83c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.275126] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 750.275312] env[63418]: DEBUG nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 750.275468] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.291674] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.401615] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524aae15-5d2b-7eab-3ebe-e06ecaf8b126, 'name': SearchDatastore_Task, 'duration_secs': 0.014266} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.401927] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.402176] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.402403] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.402544] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.402720] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.402965] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f723c5f2-0e13-4364-b79b-913327a40092 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.410731] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.410942] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 750.411584] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61e3e70d-542c-4d64-be00-079c46144172 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.416219] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 750.416219] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52343dcc-c214-bfeb-6ba4-c080e2005e06" [ 750.416219] env[63418]: _type = "Task" [ 750.416219] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.423199] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52343dcc-c214-bfeb-6ba4-c080e2005e06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.480054] env[63418]: INFO nova.compute.manager [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Took 28.69 seconds to build instance. [ 750.580086] env[63418]: DEBUG nova.scheduler.client.report [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 750.636933] env[63418]: DEBUG nova.network.neutron [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.754986] env[63418]: DEBUG oslo_concurrency.lockutils [req-4eca2f58-6d35-4567-aef4-f490998cd238 req-719c470d-82b3-4278-ab2d-3f7b8ae8d9ac service nova] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.786783] env[63418]: DEBUG nova.network.neutron [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Updating instance_info_cache with network_info: [{"id": "b62553b9-269a-4123-95eb-1886794051ef", "address": "fa:16:3e:99:75:a4", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62553b9-26", "ovs_interfaceid": "b62553b9-269a-4123-95eb-1886794051ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.793818] env[63418]: DEBUG nova.network.neutron [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.927190] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52343dcc-c214-bfeb-6ba4-c080e2005e06, 'name': SearchDatastore_Task, 'duration_secs': 0.007601} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.927978] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-712717b1-81a5-4178-81f1-ccb8bdb4bcfe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.933017] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 750.933017] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e03ce-afaf-43c8-f5d8-681c3d3e6d6a" [ 750.933017] env[63418]: _type = "Task" [ 750.933017] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.940332] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e03ce-afaf-43c8-f5d8-681c3d3e6d6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.982013] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dca8760a-baf3-4ce6-aad9-0fddf60c7e63 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.013s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.085055] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.085705] env[63418]: ERROR nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] Traceback (most recent call last): [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self.driver.spawn(context, instance, image_meta, [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] vm_ref = self.build_virtual_machine(instance, [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.085705] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] for vif in network_info: [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] return self._sync_wrapper(fn, *args, **kwargs) [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self.wait() [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self[:] = self._gt.wait() [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] return self._exit_event.wait() [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] current.throw(*self._exc) [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.086102] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] result = function(*args, **kwargs) [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] return func(*args, **kwargs) [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] raise e [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] nwinfo = self.network_api.allocate_for_instance( [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] created_port_ids = self._update_ports_for_instance( [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] with excutils.save_and_reraise_exception(): [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] self.force_reraise() [ 751.086554] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] raise self.value [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] updated_port = self._update_port( [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] _ensure_no_port_binding_failure(port) [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] raise exception.PortBindingFailed(port_id=port['id']) [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] nova.exception.PortBindingFailed: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. [ 751.086997] env[63418]: ERROR nova.compute.manager [instance: 702be472-a4bc-4089-94de-4355ee00684a] [ 751.086997] env[63418]: DEBUG nova.compute.utils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.087748] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.140s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.089260] env[63418]: INFO nova.compute.claims [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.093155] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Build of instance 702be472-a4bc-4089-94de-4355ee00684a was re-scheduled: Binding failed for port 1f437bb7-273f-4f91-b781-5c6b845c74cf, please check neutron logs for more information. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 751.093591] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 751.093875] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquiring lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.094039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Acquired lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.094201] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.290398] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-c4cb1770-b3c7-4f6b-ba11-85310f2175a0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.290730] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Instance network_info: |[{"id": "b62553b9-269a-4123-95eb-1886794051ef", "address": "fa:16:3e:99:75:a4", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62553b9-26", "ovs_interfaceid": "b62553b9-269a-4123-95eb-1886794051ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 751.291254] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:75:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b62553b9-269a-4123-95eb-1886794051ef', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.299015] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating folder: Project (c6d52f1fceb24234a8d967038b43c857). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.299556] env[63418]: INFO nova.compute.manager [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] [instance: e7a4d56b-f785-44b4-acb3-6390a4bad83c] Took 1.02 seconds to deallocate network for instance. [ 751.302132] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fecc89a6-429c-4aba-acc5-15755dd050af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.311962] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created folder: Project (c6d52f1fceb24234a8d967038b43c857) in parent group-v268354. [ 751.312173] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating folder: Instances. Parent ref: group-v268395. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.312410] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ae755bc-a349-44d7-b76c-41f6a98455e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.322832] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created folder: Instances in parent group-v268395. [ 751.322832] env[63418]: DEBUG oslo.service.loopingcall [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.322832] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.322832] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-773982be-aa39-442f-9ab4-924d5ab23728 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.352784] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.352784] env[63418]: value = "task-1244770" [ 751.352784] env[63418]: _type = "Task" [ 751.352784] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.362793] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244770, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.446390] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e03ce-afaf-43c8-f5d8-681c3d3e6d6a, 'name': SearchDatastore_Task, 'duration_secs': 0.008578} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.446726] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.447035] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 751.447309] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b8486a8-7b53-4d3e-acbf-07d673751ea8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.457235] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 751.457235] env[63418]: value = "task-1244771" [ 751.457235] env[63418]: _type = "Task" [ 751.457235] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.465727] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.484904] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 751.619283] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.713532] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.867374] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244770, 'name': CreateVM_Task, 'duration_secs': 0.505218} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.867979] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 751.868707] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.868957] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.869220] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 751.869489] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ecf8add-0794-4349-a7c4-25f6b20ee690 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.877499] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 751.877499] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ce1e19-978c-573a-017f-7ae32da0dc6c" [ 751.877499] env[63418]: _type = "Task" [ 751.877499] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.888285] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ce1e19-978c-573a-017f-7ae32da0dc6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.912339] env[63418]: DEBUG nova.compute.manager [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Received event network-vif-plugged-b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 751.912339] env[63418]: DEBUG oslo_concurrency.lockutils [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] Acquiring lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.912527] env[63418]: DEBUG oslo_concurrency.lockutils [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.912845] env[63418]: DEBUG oslo_concurrency.lockutils [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.913067] env[63418]: DEBUG nova.compute.manager [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] No waiting events found dispatching network-vif-plugged-b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 751.913193] env[63418]: WARNING nova.compute.manager [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Received unexpected event network-vif-plugged-b62553b9-269a-4123-95eb-1886794051ef for instance with vm_state building and task_state spawning. [ 751.913349] env[63418]: DEBUG nova.compute.manager [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Received event network-changed-b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 751.913502] env[63418]: DEBUG nova.compute.manager [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Refreshing instance network info cache due to event network-changed-b62553b9-269a-4123-95eb-1886794051ef. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 751.913735] env[63418]: DEBUG oslo_concurrency.lockutils [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] Acquiring lock "refresh_cache-c4cb1770-b3c7-4f6b-ba11-85310f2175a0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.913907] env[63418]: DEBUG oslo_concurrency.lockutils [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] Acquired lock "refresh_cache-c4cb1770-b3c7-4f6b-ba11-85310f2175a0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.914082] env[63418]: DEBUG nova.network.neutron [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Refreshing network info cache for port b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 751.967436] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480698} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.967794] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 751.967932] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.968207] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e94a34d6-c591-4ddb-9bb8-6958b6ab1431 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.975528] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 751.975528] env[63418]: value = "task-1244772" [ 751.975528] env[63418]: _type = "Task" [ 751.975528] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.983325] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.012488] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.219176] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Releasing lock "refresh_cache-702be472-a4bc-4089-94de-4355ee00684a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.219176] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 752.219176] env[63418]: DEBUG nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 752.219176] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.233592] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.333056] env[63418]: INFO nova.scheduler.client.report [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Deleted allocations for instance e7a4d56b-f785-44b4-acb3-6390a4bad83c [ 752.391989] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ce1e19-978c-573a-017f-7ae32da0dc6c, 'name': SearchDatastore_Task, 'duration_secs': 0.026435} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.392414] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.392671] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.392902] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.393067] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.393254] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.393570] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a97c445-bcbe-4083-8604-b2c1bfb7bf5b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.436970] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.436970] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 752.436970] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ccf846c-a505-41b4-ac7a-e5de02c44b5c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.436970] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 752.436970] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f8b468-8edc-e30a-fe5f-ae2259432e88" [ 752.436970] env[63418]: _type = "Task" [ 752.436970] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.436970] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f8b468-8edc-e30a-fe5f-ae2259432e88, 'name': SearchDatastore_Task, 'duration_secs': 0.009342} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.437328] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2b5e0d-0b62-496d-9296-d43f6da39b1c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.437328] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8be7aa7-c33d-4f57-bf5b-03b915da745b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.443020] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0753a5-1940-48ba-bdcc-109be3bfd0bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.445143] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 752.445143] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5217c7c4-c99f-090b-807e-4f3eaa6c4a99" [ 752.445143] env[63418]: _type = "Task" [ 752.445143] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.485273] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e769771-13c8-4a30-bdc9-1a5159b9bc52 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.492865] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5217c7c4-c99f-090b-807e-4f3eaa6c4a99, 'name': SearchDatastore_Task, 'duration_secs': 0.009149} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.494575] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.494884] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] c4cb1770-b3c7-4f6b-ba11-85310f2175a0/c4cb1770-b3c7-4f6b-ba11-85310f2175a0.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 752.495158] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c5578ec-2d94-45b7-b156-6c962e89759f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.501861] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06852} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.502226] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.502772] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae44edd0-a61d-43e1-9ed5-895312afb16c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.507395] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 752.507395] env[63418]: value = "task-1244773" [ 752.507395] env[63418]: _type = "Task" [ 752.507395] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.522375] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd933519-c0cb-4534-884d-4e28fcc1390a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.535346] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.535673] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1bcad56-742c-4faf-9f91-0e076b1644b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.557594] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244773, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.566124] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 752.566124] env[63418]: value = "task-1244774" [ 752.566124] env[63418]: _type = "Task" [ 752.566124] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.566605] env[63418]: DEBUG nova.compute.provider_tree [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.576123] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.736959] env[63418]: DEBUG nova.network.neutron [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.843917] env[63418]: DEBUG oslo_concurrency.lockutils [None req-557d628e-bea6-456d-98f4-1b6cc0c1dd42 tempest-ServerRescueNegativeTestJSON-1391924026 tempest-ServerRescueNegativeTestJSON-1391924026-project-member] Lock "e7a4d56b-f785-44b4-acb3-6390a4bad83c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.719s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.018703] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244773, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484696} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.018973] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] c4cb1770-b3c7-4f6b-ba11-85310f2175a0/c4cb1770-b3c7-4f6b-ba11-85310f2175a0.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 753.019181] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.019422] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5f3e4cc-c5f2-4d64-be33-73a5c8f5c392 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.025956] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 753.025956] env[63418]: value = "task-1244775" [ 753.025956] env[63418]: _type = "Task" [ 753.025956] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.033943] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.052948] env[63418]: DEBUG nova.network.neutron [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Updated VIF entry in instance network info cache for port b62553b9-269a-4123-95eb-1886794051ef. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 753.053388] env[63418]: DEBUG nova.network.neutron [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Updating instance_info_cache with network_info: [{"id": "b62553b9-269a-4123-95eb-1886794051ef", "address": "fa:16:3e:99:75:a4", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62553b9-26", "ovs_interfaceid": "b62553b9-269a-4123-95eb-1886794051ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.080330] env[63418]: DEBUG nova.scheduler.client.report [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 753.083350] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.240282] env[63418]: INFO nova.compute.manager [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] [instance: 702be472-a4bc-4089-94de-4355ee00684a] Took 1.02 seconds to deallocate network for instance. [ 753.347604] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 753.537431] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067021} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.537431] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.538134] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d15e447-b9c5-44ee-a048-e7a7f216701a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.560080] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] c4cb1770-b3c7-4f6b-ba11-85310f2175a0/c4cb1770-b3c7-4f6b-ba11-85310f2175a0.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.560677] env[63418]: DEBUG oslo_concurrency.lockutils [req-44cd4d64-1b1e-4aac-86ad-2eb846243ff9 req-8b8af21d-440e-41e1-b900-bdcbd49fcb42 service nova] Releasing lock "refresh_cache-c4cb1770-b3c7-4f6b-ba11-85310f2175a0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.561058] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa4e1154-dfd1-4e07-b5d1-4eb98feb10a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.584466] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.585248] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 753.587843] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244774, 'name': ReconfigVM_Task, 'duration_secs': 0.575029} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.589397] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.776s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.590816] env[63418]: INFO nova.compute.claims [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.593515] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfigured VM instance instance-00000031 to attach disk [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.593958] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 753.593958] env[63418]: value = "task-1244776" [ 753.593958] env[63418]: _type = "Task" [ 753.593958] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.596105] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6264918b-7b6e-4afd-bfbb-896986c80a7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.608533] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244776, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.610260] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 753.610260] env[63418]: value = "task-1244777" [ 753.610260] env[63418]: _type = "Task" [ 753.610260] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.618263] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244777, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.869422] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.090176] env[63418]: DEBUG nova.compute.utils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.091616] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 754.091754] env[63418]: DEBUG nova.network.neutron [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.108275] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244776, 'name': ReconfigVM_Task, 'duration_secs': 0.297642} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.108570] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Reconfigured VM instance instance-00000032 to attach disk [datastore1] c4cb1770-b3c7-4f6b-ba11-85310f2175a0/c4cb1770-b3c7-4f6b-ba11-85310f2175a0.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.110036] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94224b87-7eb5-4b57-bbff-7a42a297c5a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.120823] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244777, 'name': Rename_Task, 'duration_secs': 0.143022} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.122054] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 754.122357] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 754.122357] env[63418]: value = "task-1244778" [ 754.122357] env[63418]: _type = "Task" [ 754.122357] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.122532] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1eb5a2ae-4eb1-43e4-8614-92fb738e7cf2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.134785] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244778, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.135208] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 754.135208] env[63418]: value = "task-1244779" [ 754.135208] env[63418]: _type = "Task" [ 754.135208] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.142804] env[63418]: DEBUG nova.policy [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf4b273bfdc44700be92623cb508d3d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c325dd3ec4e43488c972c9d161e18b1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.147052] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244779, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.271109] env[63418]: INFO nova.scheduler.client.report [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Deleted allocations for instance 702be472-a4bc-4089-94de-4355ee00684a [ 754.302128] env[63418]: DEBUG oslo_concurrency.lockutils [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "6b81dbe2-aa9e-4561-962b-2af167234b90" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.302620] env[63418]: DEBUG oslo_concurrency.lockutils [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.302620] env[63418]: DEBUG nova.compute.manager [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 754.303563] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ff4fec-72bf-4a7f-bc26-b213292b7200 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.311753] env[63418]: DEBUG nova.compute.manager [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 754.312458] env[63418]: DEBUG nova.objects.instance [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lazy-loading 'flavor' on Instance uuid 6b81dbe2-aa9e-4561-962b-2af167234b90 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 754.598574] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 754.635753] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244778, 'name': Rename_Task, 'duration_secs': 0.219491} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.639225] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 754.640389] env[63418]: DEBUG nova.network.neutron [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Successfully created port: 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.646083] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dacd986c-53de-4163-b919-b7f7dc95d124 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.653373] env[63418]: DEBUG oslo_vmware.api [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244779, 'name': PowerOnVM_Task, 'duration_secs': 0.507527} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.659076] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 754.659420] env[63418]: INFO nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Took 7.89 seconds to spawn the instance on the hypervisor. [ 754.659643] env[63418]: DEBUG nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 754.660020] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 754.660020] env[63418]: value = "task-1244780" [ 754.660020] env[63418]: _type = "Task" [ 754.660020] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.661365] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8b4151-fd14-4d79-9793-901bdddbb324 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.680157] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244780, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.785396] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7a892af-5832-44d4-8d93-d356bb8d655a tempest-InstanceActionsNegativeTestJSON-134480384 tempest-InstanceActionsNegativeTestJSON-134480384-project-member] Lock "702be472-a4bc-4089-94de-4355ee00684a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.010s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.960170] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6731a42d-e1bf-4325-8ae4-d08beca4373e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.968703] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0546d17-8d2d-4b50-bc45-efb6b387130e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.002689] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9d99ac-7eaa-4445-8ee5-36fcba30a811 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.010683] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52593704-227b-455e-b017-dc79f84f84cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.024339] env[63418]: DEBUG nova.compute.provider_tree [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.177176] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244780, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.193609] env[63418]: INFO nova.compute.manager [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Took 31.49 seconds to build instance. [ 755.288597] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 755.323056] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.323393] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec4132dc-c530-4488-ac58-ef51aef4081f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.335593] env[63418]: DEBUG oslo_vmware.api [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 755.335593] env[63418]: value = "task-1244781" [ 755.335593] env[63418]: _type = "Task" [ 755.335593] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.348981] env[63418]: DEBUG oslo_vmware.api [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.529173] env[63418]: DEBUG nova.scheduler.client.report [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 755.608121] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 755.644878] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.645167] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.645327] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.645593] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.645632] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.645777] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.646229] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.646229] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.646331] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.646459] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.646627] env[63418]: DEBUG nova.virt.hardware [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.647600] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1490b52-e461-42a1-b513-aca6214625da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.656794] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c796545-ad29-4d76-aaaa-b0bc85a8b12a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.677231] env[63418]: DEBUG oslo_vmware.api [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244780, 'name': PowerOnVM_Task, 'duration_secs': 0.880326} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.677501] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 755.677977] env[63418]: INFO nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Took 6.41 seconds to spawn the instance on the hypervisor. [ 755.677977] env[63418]: DEBUG nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 755.678824] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b486d5-9c99-4d28-b0d3-23267cee9921 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.695778] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dfed73c2-b663-43ab-9c0f-e04060095eca tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.026s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.818312] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.845348] env[63418]: DEBUG oslo_vmware.api [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244781, 'name': PowerOffVM_Task, 'duration_secs': 0.261389} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.845626] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.845822] env[63418]: DEBUG nova.compute.manager [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 755.846625] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d300b4-fda3-4f70-892d-44a32be949c9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.034810] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.035766] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 756.038278] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.333s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.042020] env[63418]: INFO nova.compute.claims [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.198348] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 756.208020] env[63418]: INFO nova.compute.manager [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Took 30.52 seconds to build instance. [ 756.238384] env[63418]: DEBUG nova.network.neutron [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Successfully updated port: 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 756.333137] env[63418]: DEBUG nova.compute.manager [req-87507dfc-17d5-471e-a0aa-c85ea91bb9cb req-ab1b43a2-cb86-4cd2-b659-8fa444a54e31 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Received event network-vif-plugged-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 756.333368] env[63418]: DEBUG oslo_concurrency.lockutils [req-87507dfc-17d5-471e-a0aa-c85ea91bb9cb req-ab1b43a2-cb86-4cd2-b659-8fa444a54e31 service nova] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.335308] env[63418]: DEBUG oslo_concurrency.lockutils [req-87507dfc-17d5-471e-a0aa-c85ea91bb9cb req-ab1b43a2-cb86-4cd2-b659-8fa444a54e31 service nova] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.335487] env[63418]: DEBUG oslo_concurrency.lockutils [req-87507dfc-17d5-471e-a0aa-c85ea91bb9cb req-ab1b43a2-cb86-4cd2-b659-8fa444a54e31 service nova] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.335662] env[63418]: DEBUG nova.compute.manager [req-87507dfc-17d5-471e-a0aa-c85ea91bb9cb req-ab1b43a2-cb86-4cd2-b659-8fa444a54e31 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] No waiting events found dispatching network-vif-plugged-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 756.335831] env[63418]: WARNING nova.compute.manager [req-87507dfc-17d5-471e-a0aa-c85ea91bb9cb req-ab1b43a2-cb86-4cd2-b659-8fa444a54e31 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Received unexpected event network-vif-plugged-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a for instance with vm_state building and task_state spawning. [ 756.360360] env[63418]: DEBUG oslo_concurrency.lockutils [None req-278522a4-2544-48d1-808f-4381d6077c86 tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.058s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.543760] env[63418]: DEBUG nova.compute.utils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.547234] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 756.547409] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.619696] env[63418]: DEBUG nova.policy [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f7d5438e44e46d6992d08ea01dff9d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c66588ee051b4fa88353f1ba04464c58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.709614] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5efd602-6d5d-4fe4-ae4e-ccd772ccb9f1 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.032s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.733899] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.743361] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.743512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.743679] env[63418]: DEBUG nova.network.neutron [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.047725] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 757.111192] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Successfully created port: b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.140719] env[63418]: DEBUG nova.compute.manager [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 757.140912] env[63418]: DEBUG nova.compute.manager [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing instance network info cache due to event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 757.141139] env[63418]: DEBUG oslo_concurrency.lockutils [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.141281] env[63418]: DEBUG oslo_concurrency.lockutils [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.141438] env[63418]: DEBUG nova.network.neutron [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.216861] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 757.336878] env[63418]: DEBUG nova.network.neutron [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.448113] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634d0fe5-6dfd-47da-995e-98fa956f0bbb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.458761] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e73a445-d6c9-49f7-bb4a-25691da6fff9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.294542] env[63418]: DEBUG nova.objects.instance [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lazy-loading 'flavor' on Instance uuid 6b81dbe2-aa9e-4561-962b-2af167234b90 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 758.303558] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75897093-475d-4f95-87fd-35dbbc060e1b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.309018] env[63418]: DEBUG oslo_concurrency.lockutils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.309018] env[63418]: DEBUG oslo_concurrency.lockutils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.315339] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945e1bd5-e986-4ea2-be0a-41f21ba6c4e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.330584] env[63418]: DEBUG nova.compute.provider_tree [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.332492] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.422052] env[63418]: DEBUG nova.network.neutron [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updating instance_info_cache with network_info: [{"id": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "address": "fa:16:3e:c9:c8:db", "network": {"id": "1c35e426-afbd-4c95-97b6-6e351aab89ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1756487440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c325dd3ec4e43488c972c9d161e18b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f91f4e7-2e", "ovs_interfaceid": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.500793] env[63418]: DEBUG nova.compute.manager [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Received event network-changed-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 758.500981] env[63418]: DEBUG nova.compute.manager [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Refreshing instance network info cache due to event network-changed-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 758.501200] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] Acquiring lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.804171] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 758.809990] env[63418]: DEBUG oslo_concurrency.lockutils [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.810693] env[63418]: DEBUG oslo_concurrency.lockutils [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquired lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.810693] env[63418]: DEBUG nova.network.neutron [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.811052] env[63418]: DEBUG nova.objects.instance [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lazy-loading 'info_cache' on Instance uuid 6b81dbe2-aa9e-4561-962b-2af167234b90 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 758.813627] env[63418]: DEBUG nova.compute.utils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.836130] env[63418]: DEBUG nova.scheduler.client.report [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 758.847276] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.847512] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.847670] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.847845] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.847983] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.849482] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.849482] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.849482] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.849482] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.849482] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.849891] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.852030] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec76a61-208a-4066-9135-6ce73d11fe90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.859610] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b26ce18-6412-4dbe-ae07-9a7645842062 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.929947] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Releasing lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.929947] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Instance network_info: |[{"id": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "address": "fa:16:3e:c9:c8:db", "network": {"id": "1c35e426-afbd-4c95-97b6-6e351aab89ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1756487440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c325dd3ec4e43488c972c9d161e18b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f91f4e7-2e", "ovs_interfaceid": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 758.930365] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] Acquired lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.930420] env[63418]: DEBUG nova.network.neutron [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Refreshing network info cache for port 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.932596] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:c8:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8459aaf-d6a8-46fb-ad14-464ac3104695', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.942026] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Creating folder: Project (9c325dd3ec4e43488c972c9d161e18b1). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.943333] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc00751c-93e6-44af-b5e7-3baa17ad3eb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.956741] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Created folder: Project (9c325dd3ec4e43488c972c9d161e18b1) in parent group-v268354. [ 758.956906] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Creating folder: Instances. Parent ref: group-v268398. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.957218] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b80ae01e-c169-4180-ac7e-b82e253614bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.967080] env[63418]: DEBUG nova.network.neutron [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updated VIF entry in instance network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 758.967565] env[63418]: DEBUG nova.network.neutron [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.969983] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Created folder: Instances in parent group-v268398. [ 758.974028] env[63418]: DEBUG oslo.service.loopingcall [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.975476] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.975476] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1ce42cd-788b-4eb4-8bb2-cb56dca3d6f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.996726] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.996726] env[63418]: value = "task-1244784" [ 758.996726] env[63418]: _type = "Task" [ 758.996726] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.004752] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244784, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.233465] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Successfully updated port: b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 759.318028] env[63418]: DEBUG nova.objects.base [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Object Instance<6b81dbe2-aa9e-4561-962b-2af167234b90> lazy-loaded attributes: flavor,info_cache {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 759.319603] env[63418]: DEBUG oslo_concurrency.lockutils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.342031] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.304s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.342700] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 759.345021] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.516s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.475332] env[63418]: DEBUG oslo_concurrency.lockutils [req-3616b10c-5819-4d6f-881d-2514d722da9d req-b7098cea-7bed-4419-89ad-818e89da34b6 service nova] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.509977] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244784, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.681916] env[63418]: DEBUG nova.network.neutron [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updated VIF entry in instance network info cache for port 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 759.682320] env[63418]: DEBUG nova.network.neutron [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updating instance_info_cache with network_info: [{"id": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "address": "fa:16:3e:c9:c8:db", "network": {"id": "1c35e426-afbd-4c95-97b6-6e351aab89ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1756487440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c325dd3ec4e43488c972c9d161e18b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f91f4e7-2e", "ovs_interfaceid": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.735762] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "refresh_cache-7615244b-fa19-4dbb-a851-604bb6fc2725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.735934] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "refresh_cache-7615244b-fa19-4dbb-a851-604bb6fc2725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.736227] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.848846] env[63418]: DEBUG nova.compute.utils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.856309] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 759.856499] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.919057] env[63418]: DEBUG nova.policy [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f7d5438e44e46d6992d08ea01dff9d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c66588ee051b4fa88353f1ba04464c58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.016298] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244784, 'name': CreateVM_Task, 'duration_secs': 0.541423} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.016505] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 760.017639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.017639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.017710] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 760.017947] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d98f1190-24c5-463f-b93f-578d02f2e60d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.028187] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 760.028187] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c6e09c-c79f-d656-645b-4aae2061690e" [ 760.028187] env[63418]: _type = "Task" [ 760.028187] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.035898] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c6e09c-c79f-d656-645b-4aae2061690e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.186844] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4ef4175-cd5e-492b-b1f6-c1f25ae0a163 req-1fc688ae-ebc0-4ac4-9a8a-f7bbea161052 service nova] Releasing lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.193594] env[63418]: DEBUG nova.network.neutron [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updating instance_info_cache with network_info: [{"id": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "address": "fa:16:3e:a2:24:98", "network": {"id": "08e4b4ad-e54e-4539-a487-ec171aa0c5c5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1672372264-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4114cfb688a4108adcde1e1bd083ecc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d15beed-d1", "ovs_interfaceid": "9d15beed-d1a9-4234-b1ab-fd8df3c8abc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.283175] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.315482] env[63418]: DEBUG oslo_concurrency.lockutils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.315638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.315861] env[63418]: INFO nova.compute.manager [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Attaching volume b377e74e-57f3-45b6-9723-6bb19cfcceed to /dev/sdb [ 760.349710] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88815a14-41c7-4dec-875f-9f2ef4eda95b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.353194] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 760.356309] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Successfully created port: 881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.364888] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82645f45-637e-4f43-b1c4-5bb53e94525b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.388428] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6b81dbe2-aa9e-4561-962b-2af167234b90 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.388727] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 17a055e4-13da-4984-ae64-af1ae9d652f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.389198] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance c0fd1999-346d-4c12-9b8c-ab7e21ec4227 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.389198] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance d76a008c-9bd9-420b-873d-4f7d7f25b8ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.389198] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance c4cb1770-b3c7-4f6b-ba11-85310f2175a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.389333] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.389471] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7615244b-fa19-4dbb-a851-604bb6fc2725 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.389746] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 0c151b0c-c383-4421-b506-b7afa95e2072 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.391586] env[63418]: DEBUG nova.virt.block_device [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Updating existing volume attachment record: 745d1314-bea5-44dc-810f-4ad19d6f624f {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 760.508310] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Updating instance_info_cache with network_info: [{"id": "b69a59a0-e760-44f9-810e-1cfba609535b", "address": "fa:16:3e:13:ad:4d", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb69a59a0-e7", "ovs_interfaceid": "b69a59a0-e760-44f9-810e-1cfba609535b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.537628] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c6e09c-c79f-d656-645b-4aae2061690e, 'name': SearchDatastore_Task, 'duration_secs': 0.011429} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.538363] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.538639] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.538935] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.539100] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.539314] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.539580] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74f95b74-b36c-4f8a-9459-0cffc7147156 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.547460] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.547639] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 760.548352] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58056eb1-c69a-419c-bd6a-2ab37103180c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.553110] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 760.553110] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52afdf6d-10d3-00b4-9983-eb22cde93b58" [ 760.553110] env[63418]: _type = "Task" [ 760.553110] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.561147] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52afdf6d-10d3-00b4-9983-eb22cde93b58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.612319] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "849552d4-0bb4-48ee-af7b-390183bd2189" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.612590] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "849552d4-0bb4-48ee-af7b-390183bd2189" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.624891] env[63418]: DEBUG nova.compute.manager [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Received event network-vif-plugged-b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 760.625148] env[63418]: DEBUG oslo_concurrency.lockutils [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] Acquiring lock "7615244b-fa19-4dbb-a851-604bb6fc2725-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.625959] env[63418]: DEBUG oslo_concurrency.lockutils [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.626462] env[63418]: DEBUG oslo_concurrency.lockutils [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.626795] env[63418]: DEBUG nova.compute.manager [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] No waiting events found dispatching network-vif-plugged-b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 760.626980] env[63418]: WARNING nova.compute.manager [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Received unexpected event network-vif-plugged-b69a59a0-e760-44f9-810e-1cfba609535b for instance with vm_state building and task_state spawning. [ 760.627186] env[63418]: DEBUG nova.compute.manager [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Received event network-changed-b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 760.627357] env[63418]: DEBUG nova.compute.manager [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Refreshing instance network info cache due to event network-changed-b69a59a0-e760-44f9-810e-1cfba609535b. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 760.627550] env[63418]: DEBUG oslo_concurrency.lockutils [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] Acquiring lock "refresh_cache-7615244b-fa19-4dbb-a851-604bb6fc2725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.702150] env[63418]: DEBUG oslo_concurrency.lockutils [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Releasing lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.895211] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance f632d71d-498f-4914-9895-8f37187a295f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.015141] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "refresh_cache-7615244b-fa19-4dbb-a851-604bb6fc2725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.015532] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Instance network_info: |[{"id": "b69a59a0-e760-44f9-810e-1cfba609535b", "address": "fa:16:3e:13:ad:4d", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb69a59a0-e7", "ovs_interfaceid": "b69a59a0-e760-44f9-810e-1cfba609535b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 761.015887] env[63418]: DEBUG oslo_concurrency.lockutils [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] Acquired lock "refresh_cache-7615244b-fa19-4dbb-a851-604bb6fc2725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.016114] env[63418]: DEBUG nova.network.neutron [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Refreshing network info cache for port b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 761.017427] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:ad:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b69a59a0-e760-44f9-810e-1cfba609535b', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.025050] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Creating folder: Project (c66588ee051b4fa88353f1ba04464c58). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.026067] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7716e09-b824-4b47-a1e7-2b4cd7f1bedb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.037643] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Created folder: Project (c66588ee051b4fa88353f1ba04464c58) in parent group-v268354. [ 761.037853] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Creating folder: Instances. Parent ref: group-v268403. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.038093] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0f5a8b3-dca2-447f-845d-cfe01726574a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.047423] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Created folder: Instances in parent group-v268403. [ 761.047651] env[63418]: DEBUG oslo.service.loopingcall [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.047833] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.048040] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d97a22a9-8f5e-49d7-b459-714fd7a31eb2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.072219] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52afdf6d-10d3-00b4-9983-eb22cde93b58, 'name': SearchDatastore_Task, 'duration_secs': 0.010013} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.073768] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.073768] env[63418]: value = "task-1244790" [ 761.073768] env[63418]: _type = "Task" [ 761.073768] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.073971] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7351815b-7c37-4340-9ced-016b212d6c63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.082694] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 761.082694] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a73dc9-80fe-3abf-3064-4889ed3b8afd" [ 761.082694] env[63418]: _type = "Task" [ 761.082694] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.085701] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244790, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.096429] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a73dc9-80fe-3abf-3064-4889ed3b8afd, 'name': SearchDatastore_Task, 'duration_secs': 0.010058} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.096692] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.096945] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171/3e7b9a60-9e4b-46e3-adf9-7337dc7d2171.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.097230] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-571413aa-6361-48e4-b34e-86a86bb4cf9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.104396] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 761.104396] env[63418]: value = "task-1244791" [ 761.104396] env[63418]: _type = "Task" [ 761.104396] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.112287] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.364607] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 761.393236] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.393492] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.393648] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.393873] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.394044] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.394198] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.394407] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.394562] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.394732] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.394924] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.395121] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.396028] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f46cabf-6d86-4672-90fc-ef34de9778b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.399233] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance e3dd6d5e-2b80-4724-9826-6ebfb8181c3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.406229] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4a0b65-68c2-4a6f-a57c-adb8152650a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.585892] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244790, 'name': CreateVM_Task, 'duration_secs': 0.322475} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.586134] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.586751] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.586915] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.587296] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.587554] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f1dec60-c7d2-4a31-a7a7-2e93fc19d583 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.592576] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 761.592576] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d451ba-dd4f-8846-d260-58647f6510ce" [ 761.592576] env[63418]: _type = "Task" [ 761.592576] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.600322] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d451ba-dd4f-8846-d260-58647f6510ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.615228] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244791, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478568} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.615477] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171/3e7b9a60-9e4b-46e3-adf9-7337dc7d2171.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 761.615691] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.615938] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-839f4537-9772-4f56-aaa9-bb6141bb4a6e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.622130] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 761.622130] env[63418]: value = "task-1244792" [ 761.622130] env[63418]: _type = "Task" [ 761.622130] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.632031] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244792, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.708709] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.709040] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbdc845e-ecd4-422e-84de-b6d8308359d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.717996] env[63418]: DEBUG oslo_vmware.api [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 761.717996] env[63418]: value = "task-1244793" [ 761.717996] env[63418]: _type = "Task" [ 761.717996] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.732379] env[63418]: DEBUG oslo_vmware.api [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.806899] env[63418]: DEBUG nova.network.neutron [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Updated VIF entry in instance network info cache for port b69a59a0-e760-44f9-810e-1cfba609535b. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 761.807300] env[63418]: DEBUG nova.network.neutron [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Updating instance_info_cache with network_info: [{"id": "b69a59a0-e760-44f9-810e-1cfba609535b", "address": "fa:16:3e:13:ad:4d", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb69a59a0-e7", "ovs_interfaceid": "b69a59a0-e760-44f9-810e-1cfba609535b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.903259] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.999215] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Successfully updated port: 881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.102565] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d451ba-dd4f-8846-d260-58647f6510ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.102665] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.102866] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.103124] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.103269] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.103448] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.103724] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e16d38aa-7375-425f-a2f0-dad240865040 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.116918] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.116918] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.116918] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3161f88-e966-4b23-a5cd-2d73e8e9ccd1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.121970] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 762.121970] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f910f0-8b30-6542-43f6-fb65975c8680" [ 762.121970] env[63418]: _type = "Task" [ 762.121970] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.132630] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f910f0-8b30-6542-43f6-fb65975c8680, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.135876] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244792, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060084} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.136154] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.137165] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928d5213-2177-4ffb-bdb2-54a4b16332b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.162125] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171/3e7b9a60-9e4b-46e3-adf9-7337dc7d2171.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.162392] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-142bd8e7-f33c-41ac-baf4-3daa0df46192 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.180756] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 762.180756] env[63418]: value = "task-1244794" [ 762.180756] env[63418]: _type = "Task" [ 762.180756] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.188299] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244794, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.228098] env[63418]: DEBUG oslo_vmware.api [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244793, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.310682] env[63418]: DEBUG oslo_concurrency.lockutils [req-d62a4d5b-22d7-4ad8-a63c-094b3bafdd3f req-64fb6f38-7ca9-4d18-b29c-126022c6d554 service nova] Releasing lock "refresh_cache-7615244b-fa19-4dbb-a851-604bb6fc2725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.407166] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.502767] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "refresh_cache-0c151b0c-c383-4421-b506-b7afa95e2072" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.502922] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "refresh_cache-0c151b0c-c383-4421-b506-b7afa95e2072" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.503088] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.632821] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f910f0-8b30-6542-43f6-fb65975c8680, 'name': SearchDatastore_Task, 'duration_secs': 0.018709} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.633511] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c26c065-cccc-4d3a-8f6a-041accaca10a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.638953] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 762.638953] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528cccbe-a2cc-6d1f-732e-ddeff52ffadf" [ 762.638953] env[63418]: _type = "Task" [ 762.638953] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.646608] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528cccbe-a2cc-6d1f-732e-ddeff52ffadf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.689860] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244794, 'name': ReconfigVM_Task, 'duration_secs': 0.295743} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.690185] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171/3e7b9a60-9e4b-46e3-adf9-7337dc7d2171.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.690769] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3d6067f-e064-4c9f-87ef-a9da89d13f01 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.697688] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 762.697688] env[63418]: value = "task-1244795" [ 762.697688] env[63418]: _type = "Task" [ 762.697688] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.705725] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244795, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.728487] env[63418]: DEBUG oslo_vmware.api [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244793, 'name': PowerOnVM_Task, 'duration_secs': 0.776806} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.728760] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.728986] env[63418]: DEBUG nova.compute.manager [None req-852c8b73-5be8-4f74-9acb-624cf0d58acf tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 762.729838] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116a92d3-1777-4770-ab62-3f0ad009b565 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.740183] env[63418]: DEBUG nova.compute.manager [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Received event network-vif-plugged-881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 762.740400] env[63418]: DEBUG oslo_concurrency.lockutils [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] Acquiring lock "0c151b0c-c383-4421-b506-b7afa95e2072-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.740620] env[63418]: DEBUG oslo_concurrency.lockutils [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] Lock "0c151b0c-c383-4421-b506-b7afa95e2072-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.740834] env[63418]: DEBUG oslo_concurrency.lockutils [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] Lock "0c151b0c-c383-4421-b506-b7afa95e2072-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.740985] env[63418]: DEBUG nova.compute.manager [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] No waiting events found dispatching network-vif-plugged-881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 762.741193] env[63418]: WARNING nova.compute.manager [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Received unexpected event network-vif-plugged-881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 for instance with vm_state building and task_state spawning. [ 762.741345] env[63418]: DEBUG nova.compute.manager [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Received event network-changed-881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 762.741508] env[63418]: DEBUG nova.compute.manager [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Refreshing instance network info cache due to event network-changed-881c7ce8-24ab-4025-90bb-9d5c0bc92bc6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 762.741684] env[63418]: DEBUG oslo_concurrency.lockutils [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] Acquiring lock "refresh_cache-0c151b0c-c383-4421-b506-b7afa95e2072" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.911050] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 18327b62-d022-45e5-9ab0-71791491e61f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.035535] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.151705] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528cccbe-a2cc-6d1f-732e-ddeff52ffadf, 'name': SearchDatastore_Task, 'duration_secs': 0.06013} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.151995] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.152293] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 7615244b-fa19-4dbb-a851-604bb6fc2725/7615244b-fa19-4dbb-a851-604bb6fc2725.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 763.152548] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31a6fbcc-5d35-45b4-a798-b26e886bc79d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.162284] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 763.162284] env[63418]: value = "task-1244797" [ 763.162284] env[63418]: _type = "Task" [ 763.162284] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.171410] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244797, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.185953] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Updating instance_info_cache with network_info: [{"id": "881c7ce8-24ab-4025-90bb-9d5c0bc92bc6", "address": "fa:16:3e:3d:4c:2e", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap881c7ce8-24", "ovs_interfaceid": "881c7ce8-24ab-4025-90bb-9d5c0bc92bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.211177] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244795, 'name': Rename_Task, 'duration_secs': 0.143513} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.211779] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 763.212067] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-056fc45f-97d2-46f8-b3b0-527eab20d129 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.219101] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 763.219101] env[63418]: value = "task-1244798" [ 763.219101] env[63418]: _type = "Task" [ 763.219101] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.231459] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.413222] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.674346] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244797, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507925} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.674643] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 7615244b-fa19-4dbb-a851-604bb6fc2725/7615244b-fa19-4dbb-a851-604bb6fc2725.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 763.674891] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.675200] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d621b3e5-3c15-475e-bd19-ca7dc7b81fd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.681058] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 763.681058] env[63418]: value = "task-1244799" [ 763.681058] env[63418]: _type = "Task" [ 763.681058] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.690815] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "refresh_cache-0c151b0c-c383-4421-b506-b7afa95e2072" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.690815] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Instance network_info: |[{"id": "881c7ce8-24ab-4025-90bb-9d5c0bc92bc6", "address": "fa:16:3e:3d:4c:2e", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap881c7ce8-24", "ovs_interfaceid": "881c7ce8-24ab-4025-90bb-9d5c0bc92bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 763.691077] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244799, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.691077] env[63418]: DEBUG oslo_concurrency.lockutils [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] Acquired lock "refresh_cache-0c151b0c-c383-4421-b506-b7afa95e2072" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.691077] env[63418]: DEBUG nova.network.neutron [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Refreshing network info cache for port 881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.692251] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:4c:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '881c7ce8-24ab-4025-90bb-9d5c0bc92bc6', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.700763] env[63418]: DEBUG oslo.service.loopingcall [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.703907] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 763.704503] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efe0e6a0-37cb-4639-9edb-a77ffd3af5da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.731958] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244798, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.733206] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.733206] env[63418]: value = "task-1244800" [ 763.733206] env[63418]: _type = "Task" [ 763.733206] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.743802] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244800, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.916504] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 27290558-6c58-414d-bd53-c4a686890721 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.973225] env[63418]: DEBUG nova.network.neutron [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Updated VIF entry in instance network info cache for port 881c7ce8-24ab-4025-90bb-9d5c0bc92bc6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 763.973590] env[63418]: DEBUG nova.network.neutron [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Updating instance_info_cache with network_info: [{"id": "881c7ce8-24ab-4025-90bb-9d5c0bc92bc6", "address": "fa:16:3e:3d:4c:2e", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap881c7ce8-24", "ovs_interfaceid": "881c7ce8-24ab-4025-90bb-9d5c0bc92bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.192637] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244799, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.153718} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.192927] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.193749] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ed9a08-ab86-4289-bdb1-91bdd3dc4f93 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.215453] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 7615244b-fa19-4dbb-a851-604bb6fc2725/7615244b-fa19-4dbb-a851-604bb6fc2725.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.215742] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71a241b8-1811-4d99-ab33-5cf4cf713a2a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.240249] env[63418]: DEBUG oslo_vmware.api [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244798, 'name': PowerOnVM_Task, 'duration_secs': 0.704904} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.240535] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 764.240535] env[63418]: value = "task-1244801" [ 764.240535] env[63418]: _type = "Task" [ 764.240535] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.241144] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 764.241366] env[63418]: INFO nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Took 8.63 seconds to spawn the instance on the hypervisor. [ 764.241541] env[63418]: DEBUG nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 764.242307] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad79e04-724b-4ee6-8260-787ad15f9569 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.250099] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244800, 'name': CreateVM_Task, 'duration_secs': 0.474382} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.250807] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 764.251465] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.251634] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.251947] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 764.252191] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50b06487-337b-450f-86a4-34a12898e8c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.257892] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.266761] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 764.266761] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bdc25a-620f-ce23-99ef-8586d4cd84da" [ 764.266761] env[63418]: _type = "Task" [ 764.266761] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.274978] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bdc25a-620f-ce23-99ef-8586d4cd84da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.419970] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 764.476864] env[63418]: DEBUG oslo_concurrency.lockutils [req-eafd65a4-0359-44a3-8698-08b7d327b7b1 req-07073511-d1eb-4cb2-a9dc-630d9e378199 service nova] Releasing lock "refresh_cache-0c151b0c-c383-4421-b506-b7afa95e2072" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.753695] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.773463] env[63418]: INFO nova.compute.manager [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Took 31.84 seconds to build instance. [ 764.782190] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bdc25a-620f-ce23-99ef-8586d4cd84da, 'name': SearchDatastore_Task, 'duration_secs': 0.024733} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.782190] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.782190] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 764.782190] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.782329] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.782329] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 764.782329] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83c332d4-0b71-453d-98fa-831990fa5332 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.793709] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 764.794934] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 764.795230] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c99dbf0-4894-45d2-8c1b-2220322b5926 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.801302] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 764.801302] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a35d9b-4741-87cf-e3d2-c3ef9d19ee1f" [ 764.801302] env[63418]: _type = "Task" [ 764.801302] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.808449] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a35d9b-4741-87cf-e3d2-c3ef9d19ee1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.923892] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance eaa13276-1fb1-47e2-ad1f-445bc9f4c98f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 764.943415] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 764.943659] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268402', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'name': 'volume-b377e74e-57f3-45b6-9723-6bb19cfcceed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4cb1770-b3c7-4f6b-ba11-85310f2175a0', 'attached_at': '', 'detached_at': '', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'serial': 'b377e74e-57f3-45b6-9723-6bb19cfcceed'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 764.944872] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31c6b88-dc4e-4c3f-8cd4-9460485d2d27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.963103] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8c2b88-bb8e-4460-86a4-080825e809ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.990546] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] volume-b377e74e-57f3-45b6-9723-6bb19cfcceed/volume-b377e74e-57f3-45b6-9723-6bb19cfcceed.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.991012] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99d3fd0f-5af6-4073-a261-eee86e18d300 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.011320] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 765.011320] env[63418]: value = "task-1244802" [ 765.011320] env[63418]: _type = "Task" [ 765.011320] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.022658] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244802, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.262514] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244801, 'name': ReconfigVM_Task, 'duration_secs': 0.942008} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.262833] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 7615244b-fa19-4dbb-a851-604bb6fc2725/7615244b-fa19-4dbb-a851-604bb6fc2725.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.267017] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbf21a29-ab83-4993-9460-86c2323ef10d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.270895] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 765.270895] env[63418]: value = "task-1244803" [ 765.270895] env[63418]: _type = "Task" [ 765.270895] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.279185] env[63418]: DEBUG oslo_concurrency.lockutils [None req-39fc9793-15eb-4bad-a1a6-ad7610dbf5b5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.421s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.279440] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244803, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.313199] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a35d9b-4741-87cf-e3d2-c3ef9d19ee1f, 'name': SearchDatastore_Task, 'duration_secs': 0.019116} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.313199] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b398fff4-65a6-4f2d-8f7f-a001ee49bb6f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.317622] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 765.317622] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f4ac9c-1431-cbcf-e921-740e3e55f6fa" [ 765.317622] env[63418]: _type = "Task" [ 765.317622] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.325845] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f4ac9c-1431-cbcf-e921-740e3e55f6fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.426994] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance be1c134c-f36e-4cc7-b4ef-8f30793fb4df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 765.521226] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.783305] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244803, 'name': Rename_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.783672] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 765.831674] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f4ac9c-1431-cbcf-e921-740e3e55f6fa, 'name': SearchDatastore_Task, 'duration_secs': 0.010062} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.831962] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.832229] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0c151b0c-c383-4421-b506-b7afa95e2072/0c151b0c-c383-4421-b506-b7afa95e2072.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 765.832480] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c7646a0-b765-480f-82f5-0e499b62514b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.838911] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 765.838911] env[63418]: value = "task-1244804" [ 765.838911] env[63418]: _type = "Task" [ 765.838911] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.846864] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.930723] env[63418]: DEBUG nova.compute.manager [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Received event network-changed-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 765.930983] env[63418]: DEBUG nova.compute.manager [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Refreshing instance network info cache due to event network-changed-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 765.931267] env[63418]: DEBUG oslo_concurrency.lockutils [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] Acquiring lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.931426] env[63418]: DEBUG oslo_concurrency.lockutils [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] Acquired lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.931607] env[63418]: DEBUG nova.network.neutron [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Refreshing network info cache for port 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.933824] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7b0c70aa-e2bc-4131-97b4-4e53a378940a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.021438] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244802, 'name': ReconfigVM_Task, 'duration_secs': 0.789823} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.021802] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Reconfigured VM instance instance-00000032 to attach disk [datastore2] volume-b377e74e-57f3-45b6-9723-6bb19cfcceed/volume-b377e74e-57f3-45b6-9723-6bb19cfcceed.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.026637] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab0b6265-26d3-45e2-8762-34addabfcb89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.043864] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 766.043864] env[63418]: value = "task-1244805" [ 766.043864] env[63418]: _type = "Task" [ 766.043864] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.053520] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244805, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.273485] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.273951] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.274336] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.274677] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.274999] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.285266] env[63418]: INFO nova.compute.manager [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Terminating instance [ 766.297592] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244803, 'name': Rename_Task, 'duration_secs': 0.62308} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.302133] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 766.303654] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb37bc9f-6960-47c6-aeac-9ddcff701c87 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.313021] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 766.313021] env[63418]: value = "task-1244806" [ 766.313021] env[63418]: _type = "Task" [ 766.313021] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.320355] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.328614] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244806, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.353615] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244804, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.436499] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance fa4c0315-1ef5-4491-ab18-d49563b778fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.553928] env[63418]: DEBUG oslo_vmware.api [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244805, 'name': ReconfigVM_Task, 'duration_secs': 0.198987} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.554325] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268402', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'name': 'volume-b377e74e-57f3-45b6-9723-6bb19cfcceed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4cb1770-b3c7-4f6b-ba11-85310f2175a0', 'attached_at': '', 'detached_at': '', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'serial': 'b377e74e-57f3-45b6-9723-6bb19cfcceed'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 766.689274] env[63418]: DEBUG nova.network.neutron [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updated VIF entry in instance network info cache for port 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 766.689663] env[63418]: DEBUG nova.network.neutron [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updating instance_info_cache with network_info: [{"id": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "address": "fa:16:3e:c9:c8:db", "network": {"id": "1c35e426-afbd-4c95-97b6-6e351aab89ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1756487440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c325dd3ec4e43488c972c9d161e18b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f91f4e7-2e", "ovs_interfaceid": "8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.793332] env[63418]: DEBUG nova.compute.manager [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 766.793567] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.794572] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a83e29-a939-437c-945f-5ec403694fd7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.803963] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 766.804236] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ddc6f4f-5d56-4559-bbe5-26bd0a91428c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.810024] env[63418]: DEBUG oslo_vmware.api [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 766.810024] env[63418]: value = "task-1244807" [ 766.810024] env[63418]: _type = "Task" [ 766.810024] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.820528] env[63418]: DEBUG oslo_vmware.api [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244807, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.825380] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244806, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.849839] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533283} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.850101] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0c151b0c-c383-4421-b506-b7afa95e2072/0c151b0c-c383-4421-b506-b7afa95e2072.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 766.850310] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.850560] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea8ace61-34ce-4381-9c9b-725801c0832b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.857089] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 766.857089] env[63418]: value = "task-1244808" [ 766.857089] env[63418]: _type = "Task" [ 766.857089] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.865667] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.941751] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance f7554ae4-c7a7-4111-a830-10f9029dc074 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.193440] env[63418]: DEBUG oslo_concurrency.lockutils [req-e02a58a1-dd82-46e8-8670-105d4726d6bf req-07dcaac4-8d5f-48e2-a48d-fac3269e1134 service nova] Releasing lock "refresh_cache-3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.323083] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244806, 'name': PowerOnVM_Task, 'duration_secs': 0.985722} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.326188] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.326293] env[63418]: INFO nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Took 8.52 seconds to spawn the instance on the hypervisor. [ 767.326469] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 767.326757] env[63418]: DEBUG oslo_vmware.api [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244807, 'name': PowerOffVM_Task, 'duration_secs': 0.351089} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.327457] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5bf555-feb1-4a81-9431-8543f781d597 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.329874] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 767.330056] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 767.330485] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03ccfa23-6d74-4e82-a07c-70e1e51a8dcb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.367258] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068734} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.367521] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 767.368305] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6716c6-1de0-4893-82d2-40de08fd91af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.392260] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 0c151b0c-c383-4421-b506-b7afa95e2072/0c151b0c-c383-4421-b506-b7afa95e2072.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 767.392930] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bc25de1-bc3b-4e7a-a636-c82936246731 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.408604] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 767.408820] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 767.408995] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleting the datastore file [datastore1] c0fd1999-346d-4c12-9b8c-ab7e21ec4227 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 767.409617] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e05cf4a-7fea-436d-b2ac-8c52f4c2c102 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.414112] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 767.414112] env[63418]: value = "task-1244810" [ 767.414112] env[63418]: _type = "Task" [ 767.414112] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.418254] env[63418]: DEBUG oslo_vmware.api [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 767.418254] env[63418]: value = "task-1244811" [ 767.418254] env[63418]: _type = "Task" [ 767.418254] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.424993] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244810, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.429863] env[63418]: DEBUG oslo_vmware.api [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244811, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.445059] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.595053] env[63418]: DEBUG nova.objects.instance [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'flavor' on Instance uuid c4cb1770-b3c7-4f6b-ba11-85310f2175a0 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.847772] env[63418]: INFO nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Took 33.05 seconds to build instance. [ 767.926801] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244810, 'name': ReconfigVM_Task, 'duration_secs': 0.279643} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.926904] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 0c151b0c-c383-4421-b506-b7afa95e2072/0c151b0c-c383-4421-b506-b7afa95e2072.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.927606] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66ae2e89-d025-4ac9-af48-40a932c95d49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.932441] env[63418]: DEBUG oslo_vmware.api [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244811, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.510698} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.932982] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 767.933185] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 767.933359] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 767.933534] env[63418]: INFO nova.compute.manager [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Took 1.14 seconds to destroy the instance on the hypervisor. [ 767.933791] env[63418]: DEBUG oslo.service.loopingcall [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.933983] env[63418]: DEBUG nova.compute.manager [-] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 767.934086] env[63418]: DEBUG nova.network.neutron [-] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.936649] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 767.936649] env[63418]: value = "task-1244812" [ 767.936649] env[63418]: _type = "Task" [ 767.936649] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.952829] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 849552d4-0bb4-48ee-af7b-390183bd2189 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.952829] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 767.952829] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 767.955306] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244812, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.103995] env[63418]: DEBUG oslo_concurrency.lockutils [None req-90c36fea-0551-4e4d-aecc-d188424310e2 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.788s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.123605] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 768.147850] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 768.148032] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.161412] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 768.184294] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 768.315219] env[63418]: DEBUG nova.compute.manager [req-8b59cad3-1a20-4542-9d8c-08ce3667b814 req-1eb01eaa-d464-49e3-93d0-e93ad90b53b7 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Received event network-vif-deleted-1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 768.315422] env[63418]: INFO nova.compute.manager [req-8b59cad3-1a20-4542-9d8c-08ce3667b814 req-1eb01eaa-d464-49e3-93d0-e93ad90b53b7 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Neutron deleted interface 1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99; detaching it from the instance and deleting it from the info cache [ 768.315590] env[63418]: DEBUG nova.network.neutron [req-8b59cad3-1a20-4542-9d8c-08ce3667b814 req-1eb01eaa-d464-49e3-93d0-e93ad90b53b7 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.349719] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.151s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.451604] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244812, 'name': Rename_Task, 'duration_secs': 0.144339} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.455778] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 768.456160] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df6be9a7-eccc-4b82-9b16-0a84437e8781 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.462500] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 768.462500] env[63418]: value = "task-1244813" [ 768.462500] env[63418]: _type = "Task" [ 768.462500] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.474387] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.533540] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6696cd7f-5aed-4397-b747-c572389074fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.541066] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd09c536-601f-4fa2-abfc-9543ab48339b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.572680] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6393ca37-044a-4f85-bd4b-3388fc9d6be4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.580133] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fc5553-8c5d-48a1-a498-edc235609575 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.593860] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.743075] env[63418]: DEBUG nova.network.neutron [-] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.761680] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.762016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.762252] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.762451] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.762651] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.765282] env[63418]: INFO nova.compute.manager [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Terminating instance [ 768.818921] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b2a9128-dd2c-4bb2-b9ce-419d918605be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.828528] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cc01a0-0f67-4c5e-a0fd-2b1ee08a2cec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.856544] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 768.859389] env[63418]: DEBUG nova.compute.manager [req-8b59cad3-1a20-4542-9d8c-08ce3667b814 req-1eb01eaa-d464-49e3-93d0-e93ad90b53b7 service nova] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Detach interface failed, port_id=1cf7b4bf-b16f-4dbe-bad4-9a4c3dd4fc99, reason: Instance c0fd1999-346d-4c12-9b8c-ab7e21ec4227 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 768.972664] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244813, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.097388] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 769.246070] env[63418]: INFO nova.compute.manager [-] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Took 1.31 seconds to deallocate network for instance. [ 769.270104] env[63418]: DEBUG nova.compute.manager [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 769.270348] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 769.270636] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e82a6f06-c83b-43fa-8004-cded3b8c8245 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.279258] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 769.279258] env[63418]: value = "task-1244814" [ 769.279258] env[63418]: _type = "Task" [ 769.279258] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.287878] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.376195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.473339] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244813, 'name': PowerOnVM_Task, 'duration_secs': 0.513357} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.473620] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 769.474380] env[63418]: INFO nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Took 8.11 seconds to spawn the instance on the hypervisor. [ 769.474380] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 769.474810] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441723bc-f41c-42c6-b689-bb8adde7e71d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.603073] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 769.603272] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.258s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.603555] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.747s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.605224] env[63418]: INFO nova.compute.claims [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.607755] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 769.607896] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Cleaning up deleted instances {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11578}} [ 769.753961] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.789792] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244814, 'name': PowerOffVM_Task, 'duration_secs': 0.285081} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.790124] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 769.790331] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 769.790550] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268402', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'name': 'volume-b377e74e-57f3-45b6-9723-6bb19cfcceed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4cb1770-b3c7-4f6b-ba11-85310f2175a0', 'attached_at': '', 'detached_at': '', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'serial': 'b377e74e-57f3-45b6-9723-6bb19cfcceed'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 769.791413] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a1e0d8-4d00-4c19-a19b-5c439b09bc9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.813677] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377c7df1-6abb-46e4-b0f3-18a197892d82 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.821203] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4297d3-5c9d-41c4-9310-659084ec8554 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.842270] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0178aa26-8c1c-48d0-8c40-d8468ac43e8b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.857513] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] The volume has not been displaced from its original location: [datastore2] volume-b377e74e-57f3-45b6-9723-6bb19cfcceed/volume-b377e74e-57f3-45b6-9723-6bb19cfcceed.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 769.864434] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Reconfiguring VM instance instance-00000032 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 769.864841] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b98dbf5e-ebd1-480f-a6a8-ab8c866a266d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.891650] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 769.891650] env[63418]: value = "task-1244815" [ 769.891650] env[63418]: _type = "Task" [ 769.891650] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.902867] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244815, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.991557] env[63418]: INFO nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Took 27.30 seconds to build instance. [ 770.114426] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] There are 7 instances to clean {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11587}} [ 770.114621] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 99d00df2-6bf3-4ffe-b77d-f44ada631c8b] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 770.401967] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244815, 'name': ReconfigVM_Task, 'duration_secs': 0.235171} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.402286] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Reconfigured VM instance instance-00000032 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 770.407056] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db198591-8cad-47d1-8208-b162560db8b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.422724] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 770.422724] env[63418]: value = "task-1244816" [ 770.422724] env[63418]: _type = "Task" [ 770.422724] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.430819] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.493670] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "0c151b0c-c383-4421-b506-b7afa95e2072" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.255s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.621520] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: a8c7f192-4672-43cc-8c38-0c33ce633765] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 770.914665] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270342fa-dd1e-4ece-b868-c869ca5d200e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.922951] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1ac6ac-d714-4ba3-81b0-d1209a259a18 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.958064] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244816, 'name': ReconfigVM_Task, 'duration_secs': 0.127825} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.958564] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268402', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'name': 'volume-b377e74e-57f3-45b6-9723-6bb19cfcceed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4cb1770-b3c7-4f6b-ba11-85310f2175a0', 'attached_at': '', 'detached_at': '', 'volume_id': 'b377e74e-57f3-45b6-9723-6bb19cfcceed', 'serial': 'b377e74e-57f3-45b6-9723-6bb19cfcceed'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 770.958831] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.959593] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9707e3-11f6-401f-9c69-2771855a60c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.962498] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305f5633-53c3-4dac-894e-bc5bfa5d5616 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.971772] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1761ad1-bfc4-4d76-af6b-ad4ce661575e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.975376] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 770.975586] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d709b719-3028-4e4e-b21e-1b73dfd15652 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.986668] env[63418]: DEBUG nova.compute.provider_tree [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.996879] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 771.125116] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: ed637299-7e69-43d7-85f9-8fcbd6e90dec] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 771.489479] env[63418]: DEBUG nova.scheduler.client.report [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 771.515975] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.628156] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 01deeb21-0e27-497f-9b85-c85949a3533d] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 771.995108] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.995469] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 771.998163] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.999825] env[63418]: INFO nova.compute.claims [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.132685] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: f185a348-e91d-48d1-970b-473cc253cfdf] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 772.375929] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 772.376250] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 772.376445] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleting the datastore file [datastore1] c4cb1770-b3c7-4f6b-ba11-85310f2175a0 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 772.376725] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-483d08a1-752b-4823-995d-68964c9cfcb9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.383364] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 772.383364] env[63418]: value = "task-1244818" [ 772.383364] env[63418]: _type = "Task" [ 772.383364] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.390649] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.506888] env[63418]: DEBUG nova.compute.utils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.507724] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 772.507916] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.557586] env[63418]: DEBUG nova.policy [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f7d5438e44e46d6992d08ea01dff9d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c66588ee051b4fa88353f1ba04464c58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.635665] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7e43d259-f361-43d8-8f03-72b303680478] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 772.808321] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Successfully created port: 02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.894448] env[63418]: DEBUG oslo_vmware.api [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1244818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37699} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.894720] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.894914] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 772.895131] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.895362] env[63418]: INFO nova.compute.manager [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Took 3.62 seconds to destroy the instance on the hypervisor. [ 772.895531] env[63418]: DEBUG oslo.service.loopingcall [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.895768] env[63418]: DEBUG nova.compute.manager [-] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 772.895830] env[63418]: DEBUG nova.network.neutron [-] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.012608] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 773.139977] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: ba67658a-668e-4fca-aefe-e838f7b05e2a] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 773.321127] env[63418]: DEBUG nova.compute.manager [req-782169f1-7235-443a-9765-9b45916d0bf1 req-ce3c2223-3ad6-46a3-821f-33548730c197 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Received event network-vif-deleted-b62553b9-269a-4123-95eb-1886794051ef {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 773.321127] env[63418]: INFO nova.compute.manager [req-782169f1-7235-443a-9765-9b45916d0bf1 req-ce3c2223-3ad6-46a3-821f-33548730c197 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Neutron deleted interface b62553b9-269a-4123-95eb-1886794051ef; detaching it from the instance and deleting it from the info cache [ 773.321127] env[63418]: DEBUG nova.network.neutron [req-782169f1-7235-443a-9765-9b45916d0bf1 req-ce3c2223-3ad6-46a3-821f-33548730c197 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.373046] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d614c4d-3324-492a-84a9-85629c650671 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.380442] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f02a1f-7b04-4964-856e-d16721b4cba9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.410331] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b5aab3-82c2-42fb-8d50-b9bbb32bab16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.418833] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e867d6-3e59-4203-a78a-bf6ed9a5c89c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.432540] env[63418]: DEBUG nova.compute.provider_tree [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.646275] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.646543] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Cleaning up deleted instances with incomplete migration {{(pid=63418) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11616}} [ 773.770912] env[63418]: DEBUG nova.network.neutron [-] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.824388] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87f47b4f-fb5c-4ec5-a91c-9069366953a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.833882] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9d6aa5-4502-4c6b-8461-7cb40455968a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.859594] env[63418]: DEBUG nova.compute.manager [req-782169f1-7235-443a-9765-9b45916d0bf1 req-ce3c2223-3ad6-46a3-821f-33548730c197 service nova] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Detach interface failed, port_id=b62553b9-269a-4123-95eb-1886794051ef, reason: Instance c4cb1770-b3c7-4f6b-ba11-85310f2175a0 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 773.935467] env[63418]: DEBUG nova.scheduler.client.report [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 774.026620] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 774.053301] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.053556] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.053800] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.054015] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.054409] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.054409] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.054498] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.054647] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.054831] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.054965] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.055152] env[63418]: DEBUG nova.virt.hardware [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.055989] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e297ade-2ee0-4227-bf83-fdf81c736b7d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.063939] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719a7dca-7396-423e-bb8c-d94937a7a517 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.149101] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.273602] env[63418]: INFO nova.compute.manager [-] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Took 1.38 seconds to deallocate network for instance. [ 774.437234] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Successfully updated port: 02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.442027] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.442456] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 774.449018] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.433s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.449018] env[63418]: INFO nova.compute.claims [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.819567] env[63418]: INFO nova.compute.manager [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Took 0.55 seconds to detach 1 volumes for instance. [ 774.939713] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "refresh_cache-f632d71d-498f-4914-9895-8f37187a295f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.939864] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "refresh_cache-f632d71d-498f-4914-9895-8f37187a295f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.940016] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.952712] env[63418]: DEBUG nova.compute.utils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.954494] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 774.954659] env[63418]: DEBUG nova.network.neutron [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.994736] env[63418]: DEBUG nova.policy [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96d4e9330db64768964d12af5d692019', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '126d2db603334c8788951d3d3acb905e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.244347] env[63418]: DEBUG nova.network.neutron [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Successfully created port: a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.326020] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.391047] env[63418]: DEBUG nova.compute.manager [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Received event network-vif-plugged-02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 775.391380] env[63418]: DEBUG oslo_concurrency.lockutils [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] Acquiring lock "f632d71d-498f-4914-9895-8f37187a295f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.391928] env[63418]: DEBUG oslo_concurrency.lockutils [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] Lock "f632d71d-498f-4914-9895-8f37187a295f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.395790] env[63418]: DEBUG oslo_concurrency.lockutils [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] Lock "f632d71d-498f-4914-9895-8f37187a295f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.396061] env[63418]: DEBUG nova.compute.manager [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] No waiting events found dispatching network-vif-plugged-02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 775.396192] env[63418]: WARNING nova.compute.manager [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Received unexpected event network-vif-plugged-02aff9ac-be86-40b7-8b63-a237fd928606 for instance with vm_state building and task_state spawning. [ 775.396353] env[63418]: DEBUG nova.compute.manager [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Received event network-changed-02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 775.396505] env[63418]: DEBUG nova.compute.manager [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Refreshing instance network info cache due to event network-changed-02aff9ac-be86-40b7-8b63-a237fd928606. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 775.396670] env[63418]: DEBUG oslo_concurrency.lockutils [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] Acquiring lock "refresh_cache-f632d71d-498f-4914-9895-8f37187a295f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.458213] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 775.482690] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.638977] env[63418]: DEBUG nova.network.neutron [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Updating instance_info_cache with network_info: [{"id": "02aff9ac-be86-40b7-8b63-a237fd928606", "address": "fa:16:3e:65:1e:27", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aff9ac-be", "ovs_interfaceid": "02aff9ac-be86-40b7-8b63-a237fd928606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.844675] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f5e181-bf7b-4a51-8a54-93e897665d9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.852626] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1473ed-b957-4173-aab3-d2e57b0c2696 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.888202] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49119f5-03ac-43ab-9ee1-398c5ec1b2b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.896234] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d981e5bc-e826-4dd5-a79a-e32ff17b8b47 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.909428] env[63418]: DEBUG nova.compute.provider_tree [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.142042] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "refresh_cache-f632d71d-498f-4914-9895-8f37187a295f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.142404] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Instance network_info: |[{"id": "02aff9ac-be86-40b7-8b63-a237fd928606", "address": "fa:16:3e:65:1e:27", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aff9ac-be", "ovs_interfaceid": "02aff9ac-be86-40b7-8b63-a237fd928606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 776.142709] env[63418]: DEBUG oslo_concurrency.lockutils [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] Acquired lock "refresh_cache-f632d71d-498f-4914-9895-8f37187a295f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.142882] env[63418]: DEBUG nova.network.neutron [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Refreshing network info cache for port 02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.144079] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:1e:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02aff9ac-be86-40b7-8b63-a237fd928606', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.151930] env[63418]: DEBUG oslo.service.loopingcall [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.155116] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f632d71d-498f-4914-9895-8f37187a295f] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 776.155594] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed2a7fa6-6b88-480c-b82d-d1cb2a85da28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.175988] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.175988] env[63418]: value = "task-1244819" [ 776.175988] env[63418]: _type = "Task" [ 776.175988] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.183317] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244819, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.383034] env[63418]: DEBUG nova.network.neutron [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Updated VIF entry in instance network info cache for port 02aff9ac-be86-40b7-8b63-a237fd928606. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 776.383034] env[63418]: DEBUG nova.network.neutron [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Updating instance_info_cache with network_info: [{"id": "02aff9ac-be86-40b7-8b63-a237fd928606", "address": "fa:16:3e:65:1e:27", "network": {"id": "f1fe5645-15a5-4119-89e3-a6c05a5b3ea8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1312879018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c66588ee051b4fa88353f1ba04464c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aff9ac-be", "ovs_interfaceid": "02aff9ac-be86-40b7-8b63-a237fd928606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.412607] env[63418]: DEBUG nova.scheduler.client.report [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 776.473363] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 776.500669] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.501059] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.501247] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.501467] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.501761] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.502021] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.502381] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.502653] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.502856] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.503095] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.503996] env[63418]: DEBUG nova.virt.hardware [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.504777] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc10393f-00b8-4b5c-99e2-62662d7e4807 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.514208] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8b9355-2e2b-4d7e-925a-85e674f25594 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.686813] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244819, 'name': CreateVM_Task, 'duration_secs': 0.473995} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.686813] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f632d71d-498f-4914-9895-8f37187a295f] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 776.687148] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.687245] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.687534] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.687789] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3705e9e5-e586-42d5-9fce-1d80199b8880 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.692811] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 776.692811] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520124e2-6b73-4847-dc8a-90ab6891cbbe" [ 776.692811] env[63418]: _type = "Task" [ 776.692811] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.701381] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520124e2-6b73-4847-dc8a-90ab6891cbbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.823558] env[63418]: DEBUG nova.network.neutron [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Successfully updated port: a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.886360] env[63418]: DEBUG oslo_concurrency.lockutils [req-00a82109-b3a0-44b4-be2b-6b92b69c56f4 req-9c028ed3-8548-44ef-bfbe-b3948040444b service nova] Releasing lock "refresh_cache-f632d71d-498f-4914-9895-8f37187a295f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.918218] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.918767] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 776.921392] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.052s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.923147] env[63418]: INFO nova.compute.claims [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.203197] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520124e2-6b73-4847-dc8a-90ab6891cbbe, 'name': SearchDatastore_Task, 'duration_secs': 0.010828} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.203511] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.203773] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 777.204016] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.204172] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.204351] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.204601] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f65c4a01-d8d3-4477-8129-079006387585 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.216220] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.216463] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 777.217204] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09981301-0d74-4b31-8538-ae163d27349f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.222051] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 777.222051] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ec070e-e1eb-344f-c2da-f099b7bb4b7d" [ 777.222051] env[63418]: _type = "Task" [ 777.222051] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.230903] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ec070e-e1eb-344f-c2da-f099b7bb4b7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.325196] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.325526] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.325651] env[63418]: DEBUG nova.network.neutron [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.421952] env[63418]: DEBUG nova.compute.manager [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-vif-plugged-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 777.422481] env[63418]: DEBUG oslo_concurrency.lockutils [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] Acquiring lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.422922] env[63418]: DEBUG oslo_concurrency.lockutils [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.423077] env[63418]: DEBUG oslo_concurrency.lockutils [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.423312] env[63418]: DEBUG nova.compute.manager [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] No waiting events found dispatching network-vif-plugged-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 777.423539] env[63418]: WARNING nova.compute.manager [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received unexpected event network-vif-plugged-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 for instance with vm_state building and task_state spawning. [ 777.423766] env[63418]: DEBUG nova.compute.manager [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 777.423975] env[63418]: DEBUG nova.compute.manager [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing instance network info cache due to event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 777.424224] env[63418]: DEBUG oslo_concurrency.lockutils [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.428434] env[63418]: DEBUG nova.compute.utils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.434476] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 777.434660] env[63418]: DEBUG nova.network.neutron [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.477113] env[63418]: DEBUG nova.policy [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ef8f62bd75a4d109a2e1881408754cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11fffb9badce4abeadce3ab70dff7d58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.732463] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ec070e-e1eb-344f-c2da-f099b7bb4b7d, 'name': SearchDatastore_Task, 'duration_secs': 0.038806} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.735029] env[63418]: DEBUG nova.network.neutron [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Successfully created port: f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.737309] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21dd6564-10c3-413f-bdd2-e981e6b0a584 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.745333] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 777.745333] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52848fb8-61bd-fe79-8916-f8c68fdbe7b9" [ 777.745333] env[63418]: _type = "Task" [ 777.745333] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.751510] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52848fb8-61bd-fe79-8916-f8c68fdbe7b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.875437] env[63418]: DEBUG nova.network.neutron [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.933277] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 778.091321] env[63418]: DEBUG nova.network.neutron [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.244049] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f1d6e0-7331-4e00-8381-cd277a8197e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.257391] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bea1f9-bbd1-4036-92b7-e3cd7a994d22 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.261564] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52848fb8-61bd-fe79-8916-f8c68fdbe7b9, 'name': SearchDatastore_Task, 'duration_secs': 0.00932} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.261919] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.262195] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] f632d71d-498f-4914-9895-8f37187a295f/f632d71d-498f-4914-9895-8f37187a295f.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 778.262799] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec232109-763f-4654-8711-314d7b413e69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.290995] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697a9597-c7e3-42dd-bef2-18f14f113c4c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.294316] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 778.294316] env[63418]: value = "task-1244820" [ 778.294316] env[63418]: _type = "Task" [ 778.294316] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.304099] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e564aa28-ad18-4c10-a614-5ec1ebda5180 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.308601] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244820, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.318769] env[63418]: DEBUG nova.compute.provider_tree [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.594071] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.594626] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Instance network_info: |[{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 778.595090] env[63418]: DEBUG oslo_concurrency.lockutils [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.595385] env[63418]: DEBUG nova.network.neutron [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.597302] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:48:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.610618] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Creating folder: Project (126d2db603334c8788951d3d3acb905e). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.615599] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce312414-1112-4028-abdc-bb4416bdd6a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.629558] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Created folder: Project (126d2db603334c8788951d3d3acb905e) in parent group-v268354. [ 778.629707] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Creating folder: Instances. Parent ref: group-v268408. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.629875] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e183a903-11db-4b25-9b6b-d61929e8a6b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.640728] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Created folder: Instances in parent group-v268408. [ 778.641652] env[63418]: DEBUG oslo.service.loopingcall [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.641652] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.641652] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c50aad9b-772a-4ebb-8b45-b88caa2d23fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.665306] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.665306] env[63418]: value = "task-1244823" [ 778.665306] env[63418]: _type = "Task" [ 778.665306] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.677553] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244823, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.805118] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244820, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451724} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.805428] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] f632d71d-498f-4914-9895-8f37187a295f/f632d71d-498f-4914-9895-8f37187a295f.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 778.806183] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.806183] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12aa2ddc-d7ca-4f4a-8a0e-b9e3b3ef4f48 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.812062] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 778.812062] env[63418]: value = "task-1244824" [ 778.812062] env[63418]: _type = "Task" [ 778.812062] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.822106] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244824, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.824501] env[63418]: DEBUG nova.scheduler.client.report [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 778.877336] env[63418]: DEBUG nova.network.neutron [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updated VIF entry in instance network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 778.877882] env[63418]: DEBUG nova.network.neutron [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.950768] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 778.984221] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.984496] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.984693] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.984828] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.987317] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.987317] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.987317] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.987317] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.987317] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.987590] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.987590] env[63418]: DEBUG nova.virt.hardware [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.987811] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fd80d1-ecce-4f35-a1bc-674244b02527 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.996940] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422d1f56-4df9-4b93-99ba-a1176709e634 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.175909] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244823, 'name': CreateVM_Task, 'duration_secs': 0.335257} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.176199] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.177215] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.177490] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.177963] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.178339] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7ad6258-e07d-4c4f-8403-b4dd41444d82 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.184392] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 779.184392] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52830e77-c957-00d6-4fe3-e1409be47764" [ 779.184392] env[63418]: _type = "Task" [ 779.184392] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.192478] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52830e77-c957-00d6-4fe3-e1409be47764, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.322593] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244824, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.254623} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.322879] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.323670] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5814fa64-804b-468d-8cd1-03c8e6560f41 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.340227] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.340662] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 779.355062] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] f632d71d-498f-4914-9895-8f37187a295f/f632d71d-498f-4914-9895-8f37187a295f.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.356012] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.538s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.358585] env[63418]: INFO nova.compute.claims [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.364012] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-405e80b2-44b8-4b50-9ba4-d0970afaa483 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.395786] env[63418]: DEBUG nova.network.neutron [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Successfully updated port: f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 779.395786] env[63418]: DEBUG oslo_concurrency.lockutils [req-ada2053f-0dd9-4ab4-8c48-b9d75a95560d req-f8ba709b-69bc-4cd1-9047-23ff7754c2ef service nova] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.403396] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 779.403396] env[63418]: value = "task-1244825" [ 779.403396] env[63418]: _type = "Task" [ 779.403396] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.414815] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.454666] env[63418]: DEBUG nova.compute.manager [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Received event network-vif-plugged-f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 779.455418] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] Acquiring lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.455668] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.456908] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.456908] env[63418]: DEBUG nova.compute.manager [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] No waiting events found dispatching network-vif-plugged-f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 779.456908] env[63418]: WARNING nova.compute.manager [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Received unexpected event network-vif-plugged-f7766654-bf8e-4a36-b935-da5442b28a70 for instance with vm_state building and task_state spawning. [ 779.456908] env[63418]: DEBUG nova.compute.manager [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Received event network-changed-f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 779.456908] env[63418]: DEBUG nova.compute.manager [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Refreshing instance network info cache due to event network-changed-f7766654-bf8e-4a36-b935-da5442b28a70. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 779.457377] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] Acquiring lock "refresh_cache-1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.457377] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] Acquired lock "refresh_cache-1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.457377] env[63418]: DEBUG nova.network.neutron [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Refreshing network info cache for port f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.696019] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52830e77-c957-00d6-4fe3-e1409be47764, 'name': SearchDatastore_Task, 'duration_secs': 0.023144} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.696379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.696617] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.696852] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.697010] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.697199] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.697458] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15684fd6-1ad7-41ee-aa3f-0da28b6501a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.709865] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.710067] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.710834] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14e31eb4-8461-4be7-8a77-51fb737ae0fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.716378] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 779.716378] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256193d-7b5b-de87-5170-1c823e9dd0bd" [ 779.716378] env[63418]: _type = "Task" [ 779.716378] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.724601] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256193d-7b5b-de87-5170-1c823e9dd0bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.857538] env[63418]: DEBUG nova.compute.utils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.858987] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 779.860025] env[63418]: DEBUG nova.network.neutron [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.898271] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.903635] env[63418]: DEBUG nova.policy [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd24c725c044940b249e64e312e505b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02f73883249c44368eaaca666eacffc5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.917815] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.011233] env[63418]: DEBUG nova.network.neutron [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.148634] env[63418]: DEBUG nova.network.neutron [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.227573] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256193d-7b5b-de87-5170-1c823e9dd0bd, 'name': SearchDatastore_Task, 'duration_secs': 0.054145} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.228504] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69b6febe-8894-4063-bea5-54453b679f36 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.234551] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 780.234551] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b033c2-a69a-e5a6-dfc2-a32b726eaa62" [ 780.234551] env[63418]: _type = "Task" [ 780.234551] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.244325] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b033c2-a69a-e5a6-dfc2-a32b726eaa62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.299589] env[63418]: DEBUG nova.network.neutron [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Successfully created port: 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.365028] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 780.418823] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244825, 'name': ReconfigVM_Task, 'duration_secs': 0.89658} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.419367] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Reconfigured VM instance instance-00000036 to attach disk [datastore1] f632d71d-498f-4914-9895-8f37187a295f/f632d71d-498f-4914-9895-8f37187a295f.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.420218] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3ac1af0-a79e-4ba9-9cd5-a734191962bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.431030] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 780.431030] env[63418]: value = "task-1244826" [ 780.431030] env[63418]: _type = "Task" [ 780.431030] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.443553] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244826, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.652478] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bf6c34f-27c1-4737-85ba-c254466995cd req-3156b401-0555-4293-9369-9ac099a01e1a service nova] Releasing lock "refresh_cache-1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.653096] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.653263] env[63418]: DEBUG nova.network.neutron [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.731417] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f7b69a-7a30-425c-abaf-54cf1200e5c2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.741246] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f34687-f70e-4ba0-8a35-8f2e5291ff43 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.747533] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b033c2-a69a-e5a6-dfc2-a32b726eaa62, 'name': SearchDatastore_Task, 'duration_secs': 0.008956} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.748644] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.748933] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/e3dd6d5e-2b80-4724-9826-6ebfb8181c3b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.749207] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e08dbc3-1528-4327-bf20-12ca87cf2391 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.776392] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8203b146-5877-49dc-984a-a5a268781b4b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.780074] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 780.780074] env[63418]: value = "task-1244827" [ 780.780074] env[63418]: _type = "Task" [ 780.780074] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.786375] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9df9b86-db10-4d5d-808f-d66337c28237 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.792968] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244827, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.803046] env[63418]: DEBUG nova.compute.provider_tree [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.868911] env[63418]: INFO nova.virt.block_device [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Booting with volume bef2d2e5-3fa7-48bf-8c27-f4f988f36dce at /dev/sda [ 780.908674] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f144438-fa3f-41c0-9a13-8e694d000a54 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.917992] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd7a7b7-a755-4757-9bb0-29ef7731950b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.949512] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-479c7a75-6ae3-4e41-9161-e2e587061ee4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.952396] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244826, 'name': Rename_Task, 'duration_secs': 0.130666} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.952579] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 780.953161] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0f2622e-3e9c-4783-a11d-4212c2530230 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.960035] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2ac214-6505-41af-8d99-195d7080ddc7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.971956] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 780.971956] env[63418]: value = "task-1244828" [ 780.971956] env[63418]: _type = "Task" [ 780.971956] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.980965] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244828, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.991683] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43c5a80-8fc6-42b0-a16c-de7b71aa7a5a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.999662] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9e456f-118e-485c-b944-e29cbfd51736 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.014804] env[63418]: DEBUG nova.virt.block_device [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updating existing volume attachment record: e0858592-335e-4acc-b3b2-5cd6bd9b341c {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 781.188932] env[63418]: DEBUG nova.network.neutron [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.290845] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244827, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43145} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.290972] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/e3dd6d5e-2b80-4724-9826-6ebfb8181c3b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.291277] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.291551] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cbae981-0882-4fa8-b682-510e77a518c0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.298148] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 781.298148] env[63418]: value = "task-1244829" [ 781.298148] env[63418]: _type = "Task" [ 781.298148] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.305519] env[63418]: DEBUG nova.scheduler.client.report [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 781.312189] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244829, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.345617] env[63418]: DEBUG nova.network.neutron [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Updating instance_info_cache with network_info: [{"id": "f7766654-bf8e-4a36-b935-da5442b28a70", "address": "fa:16:3e:07:99:24", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7766654-bf", "ovs_interfaceid": "f7766654-bf8e-4a36-b935-da5442b28a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.482791] env[63418]: DEBUG oslo_vmware.api [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244828, 'name': PowerOnVM_Task, 'duration_secs': 0.469771} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.482791] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 781.482791] env[63418]: INFO nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Took 7.46 seconds to spawn the instance on the hypervisor. [ 781.483015] env[63418]: DEBUG nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 781.483672] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bff0736-18e7-437d-9304-b86b2ebeb0f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.730165] env[63418]: DEBUG nova.compute.manager [req-e7f40f1c-2529-473c-a7f4-e21ce1e587bd req-e9b3ec3c-932b-4fc3-b5c5-19eb1dcefba4 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Received event network-vif-plugged-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 781.730416] env[63418]: DEBUG oslo_concurrency.lockutils [req-e7f40f1c-2529-473c-a7f4-e21ce1e587bd req-e9b3ec3c-932b-4fc3-b5c5-19eb1dcefba4 service nova] Acquiring lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.730719] env[63418]: DEBUG oslo_concurrency.lockutils [req-e7f40f1c-2529-473c-a7f4-e21ce1e587bd req-e9b3ec3c-932b-4fc3-b5c5-19eb1dcefba4 service nova] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.730828] env[63418]: DEBUG oslo_concurrency.lockutils [req-e7f40f1c-2529-473c-a7f4-e21ce1e587bd req-e9b3ec3c-932b-4fc3-b5c5-19eb1dcefba4 service nova] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.731192] env[63418]: DEBUG nova.compute.manager [req-e7f40f1c-2529-473c-a7f4-e21ce1e587bd req-e9b3ec3c-932b-4fc3-b5c5-19eb1dcefba4 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] No waiting events found dispatching network-vif-plugged-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 781.731433] env[63418]: WARNING nova.compute.manager [req-e7f40f1c-2529-473c-a7f4-e21ce1e587bd req-e9b3ec3c-932b-4fc3-b5c5-19eb1dcefba4 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Received unexpected event network-vif-plugged-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 for instance with vm_state building and task_state block_device_mapping. [ 781.807685] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244829, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072478} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.807959] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.808720] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0c5b56-6047-46a0-a895-d551f14135d5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.822646] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.823103] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 781.835432] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/e3dd6d5e-2b80-4724-9826-6ebfb8181c3b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.835963] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.102s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.838415] env[63418]: INFO nova.compute.claims [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.840328] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-762c7921-3eef-4536-9311-d758f3e64302 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.855782] env[63418]: DEBUG nova.network.neutron [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Successfully updated port: 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.860612] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.860612] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance network_info: |[{"id": "f7766654-bf8e-4a36-b935-da5442b28a70", "address": "fa:16:3e:07:99:24", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7766654-bf", "ovs_interfaceid": "f7766654-bf8e-4a36-b935-da5442b28a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 781.860851] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:99:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7766654-bf8e-4a36-b935-da5442b28a70', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.867116] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating folder: Project (11fffb9badce4abeadce3ab70dff7d58). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.868873] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc3ea121-14cb-4f96-bd24-e6e298b09e95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.875698] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 781.875698] env[63418]: value = "task-1244830" [ 781.875698] env[63418]: _type = "Task" [ 781.875698] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.879791] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created folder: Project (11fffb9badce4abeadce3ab70dff7d58) in parent group-v268354. [ 781.879963] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating folder: Instances. Parent ref: group-v268411. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.880192] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59baf9d1-bd9c-4563-b5e9-1e8c2ca368ba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.884890] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244830, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.893126] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created folder: Instances in parent group-v268411. [ 781.893358] env[63418]: DEBUG oslo.service.loopingcall [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.893543] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 781.893740] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b64aa59-3618-426f-bacf-d0213e7d3e8a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.912318] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.912318] env[63418]: value = "task-1244833" [ 781.912318] env[63418]: _type = "Task" [ 781.912318] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.921607] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244833, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.000993] env[63418]: INFO nova.compute.manager [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Took 37.17 seconds to build instance. [ 782.337683] env[63418]: DEBUG nova.compute.utils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.339151] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 782.339316] env[63418]: DEBUG nova.network.neutron [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.359057] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.359216] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquired lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.359362] env[63418]: DEBUG nova.network.neutron [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.384819] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244830, 'name': ReconfigVM_Task, 'duration_secs': 0.27824} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.386591] env[63418]: DEBUG nova.policy [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49d2571b1af4798a5e8790e8246d114', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce61580944e04fd2ac6defdb79be636a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.389035] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Reconfigured VM instance instance-00000037 to attach disk [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/e3dd6d5e-2b80-4724-9826-6ebfb8181c3b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.389035] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-289b12ba-3935-4c81-a2ff-ca2ef6d0c10d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.394959] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 782.394959] env[63418]: value = "task-1244834" [ 782.394959] env[63418]: _type = "Task" [ 782.394959] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.403851] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244834, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.423182] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244833, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.503123] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c14b45e1-8fe5-4ce0-8b80-62faae46eff9 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "f632d71d-498f-4914-9895-8f37187a295f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.236s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.670134] env[63418]: DEBUG nova.network.neutron [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Successfully created port: 50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.842371] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 782.907528] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244834, 'name': Rename_Task, 'duration_secs': 0.154864} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.908545] env[63418]: DEBUG nova.network.neutron [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.909982] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.912533] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d39b18e-83c4-4dbc-89ad-02c21403b4e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.922542] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244833, 'name': CreateVM_Task, 'duration_secs': 0.645002} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.927683] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.928031] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 782.928031] env[63418]: value = "task-1244835" [ 782.928031] env[63418]: _type = "Task" [ 782.928031] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.930723] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.930886] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.931228] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.934388] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30214c78-3347-4b0c-8748-34ad69ddc3b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.949156] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 782.949156] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523ea490-b805-fada-723d-1ccbda787f10" [ 782.949156] env[63418]: _type = "Task" [ 782.949156] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.952799] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244835, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.960482] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523ea490-b805-fada-723d-1ccbda787f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.006167] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 783.081574] env[63418]: DEBUG nova.network.neutron [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updating instance_info_cache with network_info: [{"id": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "address": "fa:16:3e:fc:88:79", "network": {"id": "4d39fd83-617d-47b7-a1e6-603411a20c2b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1579674115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02f73883249c44368eaaca666eacffc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72cfead0-4e", "ovs_interfaceid": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.104811] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "7615244b-fa19-4dbb-a851-604bb6fc2725" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.105106] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.105317] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "7615244b-fa19-4dbb-a851-604bb6fc2725-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.105497] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.105663] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.107665] env[63418]: INFO nova.compute.manager [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Terminating instance [ 783.122214] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 783.122750] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.122967] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.123135] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.123315] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.123455] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.123594] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.123821] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.123986] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.124177] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.124339] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.124506] env[63418]: DEBUG nova.virt.hardware [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.127771] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808b9eba-9e5d-49f9-be4f-dba1ed9680da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.136155] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8e6469-8d97-4d32-831b-b10636eac372 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.186033] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96c0674-3467-4a75-ba63-d34b10e33ef9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.195216] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2bd3305-f9b6-499b-8ce8-44a3a1cb4b5e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.227226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ec774a-b57f-4321-8d24-249a0e23acd9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.234837] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d5472e-893f-4d4e-a797-59b9dc8aa58c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.250145] env[63418]: DEBUG nova.compute.provider_tree [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.441377] env[63418]: DEBUG oslo_vmware.api [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244835, 'name': PowerOnVM_Task, 'duration_secs': 0.433246} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.441722] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.441957] env[63418]: INFO nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Took 6.97 seconds to spawn the instance on the hypervisor. [ 783.442224] env[63418]: DEBUG nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 783.443029] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f300e2-dec1-4e88-b4d0-cf453f6a174e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.461980] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523ea490-b805-fada-723d-1ccbda787f10, 'name': SearchDatastore_Task, 'duration_secs': 0.010702} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.462156] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.462549] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.462667] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.463072] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.463072] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.463347] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7409d78b-da01-4e5c-b308-4ffa4349d9fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.471905] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.472125] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.472835] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-889e26c7-c860-408a-9774-0554a0da2058 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.478283] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 783.478283] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5234a08d-ae32-8eb2-9489-9f0d8d5a46b5" [ 783.478283] env[63418]: _type = "Task" [ 783.478283] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.486656] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5234a08d-ae32-8eb2-9489-9f0d8d5a46b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.525264] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.584599] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Releasing lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.584958] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance network_info: |[{"id": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "address": "fa:16:3e:fc:88:79", "network": {"id": "4d39fd83-617d-47b7-a1e6-603411a20c2b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1579674115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02f73883249c44368eaaca666eacffc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72cfead0-4e", "ovs_interfaceid": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 783.585411] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:88:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2c68e7-b690-42e2-9491-c3f9357cc66a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.593030] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Creating folder: Project (02f73883249c44368eaaca666eacffc5). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 783.593294] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63c9f343-adbd-49f2-9478-2f47b950b880 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.606657] env[63418]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 783.606813] env[63418]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63418) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 783.607177] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Folder already exists: Project (02f73883249c44368eaaca666eacffc5). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 783.607368] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Creating folder: Instances. Parent ref: group-v268379. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 783.607601] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7104116-3e40-45ec-afd5-3317124fb27a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.610994] env[63418]: DEBUG nova.compute.manager [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 783.611210] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.611953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f730d752-610a-4cc6-af59-783b74a48e3c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.616748] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Created folder: Instances in parent group-v268379. [ 783.616961] env[63418]: DEBUG oslo.service.loopingcall [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.617153] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 783.617341] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80510ef9-ee8b-4400-a004-7e5b0cf166e7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.634923] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.635552] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41728864-de4a-4db7-95ee-2c110b9f29f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.640587] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.640587] env[63418]: value = "task-1244838" [ 783.640587] env[63418]: _type = "Task" [ 783.640587] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.641811] env[63418]: DEBUG oslo_vmware.api [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 783.641811] env[63418]: value = "task-1244839" [ 783.641811] env[63418]: _type = "Task" [ 783.641811] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.652778] env[63418]: DEBUG oslo_vmware.api [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.655554] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244838, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.753998] env[63418]: DEBUG nova.scheduler.client.report [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 783.827046] env[63418]: DEBUG nova.compute.manager [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Received event network-changed-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 783.827336] env[63418]: DEBUG nova.compute.manager [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Refreshing instance network info cache due to event network-changed-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 783.827623] env[63418]: DEBUG oslo_concurrency.lockutils [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] Acquiring lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.827910] env[63418]: DEBUG oslo_concurrency.lockutils [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] Acquired lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.828186] env[63418]: DEBUG nova.network.neutron [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Refreshing network info cache for port 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 783.851200] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 783.878579] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.878874] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.879068] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.879256] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.879409] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.879572] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.879763] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.879926] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.880106] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.880353] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.880533] env[63418]: DEBUG nova.virt.hardware [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.881463] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652ba583-73db-4131-a320-787f0988357a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.891133] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42e3528-5cbf-4556-9329-0fe2526d5821 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.963212] env[63418]: INFO nova.compute.manager [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Took 37.98 seconds to build instance. [ 783.988350] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5234a08d-ae32-8eb2-9489-9f0d8d5a46b5, 'name': SearchDatastore_Task, 'duration_secs': 0.009741} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.989213] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3b135d4-b2fa-427d-a7c9-e7942a2dec20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.994935] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 783.994935] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c9bec7-a8ad-2954-157c-69a51db8a46e" [ 783.994935] env[63418]: _type = "Task" [ 783.994935] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.003238] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c9bec7-a8ad-2954-157c-69a51db8a46e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.153661] env[63418]: DEBUG oslo_vmware.api [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244839, 'name': PowerOffVM_Task, 'duration_secs': 0.250905} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.156845] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.157034] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 784.157762] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244838, 'name': CreateVM_Task, 'duration_secs': 0.367325} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.157989] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93092b64-e815-4f79-950f-057c9ee9c0be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.159409] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 784.160075] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'attachment_id': 'e0858592-335e-4acc-b3b2-5cd6bd9b341c', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268382', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'name': 'volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'serial': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce'}, 'disk_bus': None, 'boot_index': 0, 'mount_device': '/dev/sda', 'device_type': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=63418) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 784.160291] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Root volume attach. Driver type: vmdk {{(pid=63418) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 784.161480] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b97e51-317f-419f-9a26-d2a75d188b3c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.169522] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f997fc5c-1289-4463-9be9-27edfd87110d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.175556] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c18fe1-f506-4086-afa6-3b4c42d6f864 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.185991] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-6fc78548-aa44-45b5-b10f-9adf8ac27112 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.195023] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 784.195023] env[63418]: value = "task-1244841" [ 784.195023] env[63418]: _type = "Task" [ 784.195023] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.203377] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.204225] env[63418]: DEBUG nova.network.neutron [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Successfully updated port: 50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.238309] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 784.238520] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 784.238748] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleting the datastore file [datastore2] 7615244b-fa19-4dbb-a851-604bb6fc2725 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.239026] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c256432-4bd5-4d57-b645-8a1e3ffdda9f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.245746] env[63418]: DEBUG oslo_vmware.api [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 784.245746] env[63418]: value = "task-1244842" [ 784.245746] env[63418]: _type = "Task" [ 784.245746] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.254405] env[63418]: DEBUG oslo_vmware.api [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.259258] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.259777] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 784.262273] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.930s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.263693] env[63418]: INFO nova.compute.claims [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.465363] env[63418]: DEBUG oslo_concurrency.lockutils [None req-10095cc9-613d-40e8-bc91-c44be272220e tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.641s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.506310] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c9bec7-a8ad-2954-157c-69a51db8a46e, 'name': SearchDatastore_Task, 'duration_secs': 0.009626} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.506670] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.506975] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.507311] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9897fab5-f74c-43bc-8939-a61bf3e52b90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.515120] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 784.515120] env[63418]: value = "task-1244843" [ 784.515120] env[63418]: _type = "Task" [ 784.515120] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.527216] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.709421] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "refresh_cache-18327b62-d022-45e5-9ab0-71791491e61f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.709682] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquired lock "refresh_cache-18327b62-d022-45e5-9ab0-71791491e61f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.709932] env[63418]: DEBUG nova.network.neutron [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.712454] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 42%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.759260] env[63418]: DEBUG oslo_vmware.api [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134096} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.760330] env[63418]: DEBUG nova.network.neutron [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updated VIF entry in instance network info cache for port 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 784.760668] env[63418]: DEBUG nova.network.neutron [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updating instance_info_cache with network_info: [{"id": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "address": "fa:16:3e:fc:88:79", "network": {"id": "4d39fd83-617d-47b7-a1e6-603411a20c2b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1579674115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02f73883249c44368eaaca666eacffc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72cfead0-4e", "ovs_interfaceid": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.762293] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.762491] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.762677] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.762861] env[63418]: INFO nova.compute.manager [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Took 1.15 seconds to destroy the instance on the hypervisor. [ 784.763137] env[63418]: DEBUG oslo.service.loopingcall [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.763358] env[63418]: DEBUG nova.compute.manager [-] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 784.763439] env[63418]: DEBUG nova.network.neutron [-] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.767643] env[63418]: DEBUG nova.compute.utils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.774354] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 784.774354] env[63418]: DEBUG nova.network.neutron [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.857397] env[63418]: DEBUG nova.policy [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea507bad11c3406d880ba47d08a047c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd445600834dd4c7e8022349ee993f3ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.968373] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 785.032080] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490307} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.032576] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.032857] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.033291] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df0fff4e-e053-47c4-be6c-645a04984da2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.041917] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 785.041917] env[63418]: value = "task-1244844" [ 785.041917] env[63418]: _type = "Task" [ 785.041917] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.053484] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.211877] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 56%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.267040] env[63418]: DEBUG oslo_concurrency.lockutils [req-79af0a59-137a-499e-8685-186a3dced1f3 req-b35e6d81-bb26-4718-9382-6aabfe664fda service nova] Releasing lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.276376] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 785.290182] env[63418]: DEBUG nova.network.neutron [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.309473] env[63418]: INFO nova.compute.manager [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Rescuing [ 785.309802] env[63418]: DEBUG oslo_concurrency.lockutils [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.309986] env[63418]: DEBUG oslo_concurrency.lockutils [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.310255] env[63418]: DEBUG nova.network.neutron [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.319281] env[63418]: DEBUG nova.network.neutron [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Successfully created port: 75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.494143] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.562393] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081263} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.566681] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.568026] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcb8ceb-44d9-4dc5-8abe-c737ca00ec71 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.599434] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.600805] env[63418]: DEBUG nova.network.neutron [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Updating instance_info_cache with network_info: [{"id": "50971ce3-53f7-4832-8816-2252c1f95541", "address": "fa:16:3e:9a:25:f2", "network": {"id": "a8d96897-b5a2-4ad5-924a-3e001b7fadab", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2042467263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce61580944e04fd2ac6defdb79be636a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50971ce3-53", "ovs_interfaceid": "50971ce3-53f7-4832-8816-2252c1f95541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.604803] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-619120cf-50f3-46e9-9477-a4d1dfc21e1a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.623614] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Releasing lock "refresh_cache-18327b62-d022-45e5-9ab0-71791491e61f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.624154] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Instance network_info: |[{"id": "50971ce3-53f7-4832-8816-2252c1f95541", "address": "fa:16:3e:9a:25:f2", "network": {"id": "a8d96897-b5a2-4ad5-924a-3e001b7fadab", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2042467263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce61580944e04fd2ac6defdb79be636a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50971ce3-53", "ovs_interfaceid": "50971ce3-53f7-4832-8816-2252c1f95541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 785.625146] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:25:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9ee6f9-33be-4f58-8248-694024ec31d4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '50971ce3-53f7-4832-8816-2252c1f95541', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.633909] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Creating folder: Project (ce61580944e04fd2ac6defdb79be636a). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.638103] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-464a98b6-dd5d-4d40-bdf4-65e2be762638 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.644932] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 785.644932] env[63418]: value = "task-1244845" [ 785.644932] env[63418]: _type = "Task" [ 785.644932] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.658762] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Created folder: Project (ce61580944e04fd2ac6defdb79be636a) in parent group-v268354. [ 785.658979] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Creating folder: Instances. Parent ref: group-v268416. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.659294] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244845, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.659530] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-283176a0-d351-4556-894d-bec532d23b5b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.673043] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Created folder: Instances in parent group-v268416. [ 785.673430] env[63418]: DEBUG oslo.service.loopingcall [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.673678] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.673948] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d799365b-12ab-49d8-912a-44694c2dc6c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.702184] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.702184] env[63418]: value = "task-1244848" [ 785.702184] env[63418]: _type = "Task" [ 785.702184] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.712571] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 69%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.719123] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244848, 'name': CreateVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.746246] env[63418]: DEBUG nova.network.neutron [-] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.855542] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7e4654-6ddd-4518-a7ed-aeb1ba299d07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.863201] env[63418]: DEBUG nova.compute.manager [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Received event network-vif-plugged-50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 785.863322] env[63418]: DEBUG oslo_concurrency.lockutils [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] Acquiring lock "18327b62-d022-45e5-9ab0-71791491e61f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.863574] env[63418]: DEBUG oslo_concurrency.lockutils [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] Lock "18327b62-d022-45e5-9ab0-71791491e61f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.863809] env[63418]: DEBUG oslo_concurrency.lockutils [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] Lock "18327b62-d022-45e5-9ab0-71791491e61f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.864065] env[63418]: DEBUG nova.compute.manager [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] No waiting events found dispatching network-vif-plugged-50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 785.864283] env[63418]: WARNING nova.compute.manager [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Received unexpected event network-vif-plugged-50971ce3-53f7-4832-8816-2252c1f95541 for instance with vm_state building and task_state spawning. [ 785.864485] env[63418]: DEBUG nova.compute.manager [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Received event network-changed-50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 785.864679] env[63418]: DEBUG nova.compute.manager [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Refreshing instance network info cache due to event network-changed-50971ce3-53f7-4832-8816-2252c1f95541. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 785.864909] env[63418]: DEBUG oslo_concurrency.lockutils [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] Acquiring lock "refresh_cache-18327b62-d022-45e5-9ab0-71791491e61f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.865101] env[63418]: DEBUG oslo_concurrency.lockutils [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] Acquired lock "refresh_cache-18327b62-d022-45e5-9ab0-71791491e61f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.865303] env[63418]: DEBUG nova.network.neutron [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Refreshing network info cache for port 50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.870131] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5e6b8e-dfc4-4156-a0e6-db193d1ffd10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.911468] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060f0f51-0ef2-42ce-94b9-6044c06944a1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.921294] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693cc580-58a8-4870-9c5c-19cbb5a98d72 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.937264] env[63418]: DEBUG nova.compute.provider_tree [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.156684] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244845, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.183738] env[63418]: DEBUG nova.network.neutron [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.212917] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 82%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.220819] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244848, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.252721] env[63418]: INFO nova.compute.manager [-] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Took 1.49 seconds to deallocate network for instance. [ 786.290070] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 786.319947] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.320221] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.320376] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.320557] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.320710] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.320851] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.321074] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.321238] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.321408] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.321572] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.321782] env[63418]: DEBUG nova.virt.hardware [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.322716] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b034e57b-f68b-4eb1-b2af-c27e2566968e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.336866] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b725a6-3449-401a-bb49-6339e32bab98 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.442136] env[63418]: DEBUG nova.scheduler.client.report [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 786.656327] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244845, 'name': ReconfigVM_Task, 'duration_secs': 0.716052} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.656629] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.657395] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-023dce48-748b-4752-81af-3f3e16850322 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.667889] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 786.667889] env[63418]: value = "task-1244849" [ 786.667889] env[63418]: _type = "Task" [ 786.667889] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.677760] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244849, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.688128] env[63418]: DEBUG oslo_concurrency.lockutils [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.713552] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.722720] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244848, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.765170] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.815446] env[63418]: DEBUG nova.network.neutron [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Updated VIF entry in instance network info cache for port 50971ce3-53f7-4832-8816-2252c1f95541. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 786.815798] env[63418]: DEBUG nova.network.neutron [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Updating instance_info_cache with network_info: [{"id": "50971ce3-53f7-4832-8816-2252c1f95541", "address": "fa:16:3e:9a:25:f2", "network": {"id": "a8d96897-b5a2-4ad5-924a-3e001b7fadab", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2042467263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce61580944e04fd2ac6defdb79be636a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50971ce3-53", "ovs_interfaceid": "50971ce3-53f7-4832-8816-2252c1f95541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.910803] env[63418]: DEBUG nova.network.neutron [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Successfully updated port: 75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.949475] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.950061] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 786.953265] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.633s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.954833] env[63418]: INFO nova.compute.claims [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.177855] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244849, 'name': Rename_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.213948] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 97%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.219777] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244848, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.320739] env[63418]: DEBUG oslo_concurrency.lockutils [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] Releasing lock "refresh_cache-18327b62-d022-45e5-9ab0-71791491e61f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.321997] env[63418]: DEBUG nova.compute.manager [req-5438b9d4-a701-4b75-9bfd-e31ca917a54d req-d64e208d-d93a-4d89-bf19-794c2aa53e7a service nova] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Received event network-vif-deleted-b69a59a0-e760-44f9-810e-1cfba609535b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 787.415391] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "refresh_cache-0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.415668] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "refresh_cache-0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.415847] env[63418]: DEBUG nova.network.neutron [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.459677] env[63418]: DEBUG nova.compute.utils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.465903] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 787.465903] env[63418]: DEBUG nova.network.neutron [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 787.502822] env[63418]: DEBUG nova.policy [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24692bc22b0c47efba9863b48d1b5737', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88d4c2f44b5641acb3ebbf83967e1beb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.679878] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244849, 'name': Rename_Task, 'duration_secs': 0.635308} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.679878] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.679878] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e98ed9e-5e86-469b-9b46-9acfe6095e2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.687657] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 787.687657] env[63418]: value = "task-1244850" [ 787.687657] env[63418]: _type = "Task" [ 787.687657] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.696354] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.714463] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 97%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.723068] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244848, 'name': CreateVM_Task, 'duration_secs': 1.577663} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.723350] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 787.724309] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.724579] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.725052] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.725369] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac38e85e-eea0-45c7-a8d9-669ee1ec874a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.732258] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 787.732258] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aaa436-f7d9-1b9a-e158-2fcfa1489d86" [ 787.732258] env[63418]: _type = "Task" [ 787.732258] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.733017] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 787.736870] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-177cd5b4-4b43-49b0-b9a6-5c0e77ba8aff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.747033] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aaa436-f7d9-1b9a-e158-2fcfa1489d86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.748743] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 787.748743] env[63418]: value = "task-1244851" [ 787.748743] env[63418]: _type = "Task" [ 787.748743] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.761883] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.777806] env[63418]: DEBUG nova.network.neutron [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Successfully created port: f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.928202] env[63418]: DEBUG nova.compute.manager [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Received event network-vif-plugged-75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 787.928460] env[63418]: DEBUG oslo_concurrency.lockutils [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] Acquiring lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.928838] env[63418]: DEBUG oslo_concurrency.lockutils [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.929258] env[63418]: DEBUG oslo_concurrency.lockutils [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.929407] env[63418]: DEBUG nova.compute.manager [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] No waiting events found dispatching network-vif-plugged-75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 787.929587] env[63418]: WARNING nova.compute.manager [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Received unexpected event network-vif-plugged-75354cd4-70a1-4aa8-aab4-6ae9cb843f5a for instance with vm_state building and task_state spawning. [ 787.929749] env[63418]: DEBUG nova.compute.manager [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Received event network-changed-75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 787.929929] env[63418]: DEBUG nova.compute.manager [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Refreshing instance network info cache due to event network-changed-75354cd4-70a1-4aa8-aab4-6ae9cb843f5a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 787.930207] env[63418]: DEBUG oslo_concurrency.lockutils [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] Acquiring lock "refresh_cache-0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.964603] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 787.968434] env[63418]: DEBUG nova.network.neutron [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.197720] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244850, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.211615] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task} progress is 98%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.246234] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aaa436-f7d9-1b9a-e158-2fcfa1489d86, 'name': SearchDatastore_Task, 'duration_secs': 0.032255} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.246547] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.246670] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.246914] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.247195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.247286] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.247592] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b932add-d582-4cc6-877c-38c7b3a00966 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.262107] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244851, 'name': PowerOffVM_Task, 'duration_secs': 0.276929} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.262453] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.262807] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.262963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.266094] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ec87ec-c8af-4a09-a73d-0de4fd5b79f5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.266687] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a62bc0b6-d5c7-49a9-aaa2-8f3b38c7a4bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.272559] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 788.272559] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527bfe75-a144-35d8-567b-c68c016fafcc" [ 788.272559] env[63418]: _type = "Task" [ 788.272559] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.296755] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be234e9-417b-4008-8b27-cc5ca8ceea84 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.306698] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527bfe75-a144-35d8-567b-c68c016fafcc, 'name': SearchDatastore_Task, 'duration_secs': 0.011583} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.311796] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc96c99d-0284-4344-8eed-a272028f0a6d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.319702] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 788.319702] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fce02-b38e-b703-54db-b69f773c0d43" [ 788.319702] env[63418]: _type = "Task" [ 788.319702] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.330748] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fce02-b38e-b703-54db-b69f773c0d43, 'name': SearchDatastore_Task, 'duration_secs': 0.010373} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.331200] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.331617] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 18327b62-d022-45e5-9ab0-71791491e61f/18327b62-d022-45e5-9ab0-71791491e61f.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 788.331956] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70938f22-7884-43ce-a3fa-c1873f988aea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.342054] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 788.342054] env[63418]: value = "task-1244852" [ 788.342054] env[63418]: _type = "Task" [ 788.342054] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.343851] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.344221] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-085aed7e-79a3-4395-80f5-826eb52eb123 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.358260] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.360160] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 788.360160] env[63418]: value = "task-1244853" [ 788.360160] env[63418]: _type = "Task" [ 788.360160] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.366334] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab421396-f0f5-44a8-b61b-7967f4030614 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.380675] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248275eb-f4fd-4f89-9494-f64e5e765951 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.383988] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 788.385851] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.385851] env[63418]: DEBUG oslo_concurrency.lockutils [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.385851] env[63418]: DEBUG oslo_concurrency.lockutils [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.385851] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.385851] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e99ee74b-320c-41c2-bda6-0783f9507aa2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.417456] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6094b5bd-d6ad-46e4-b6aa-4c6e6de5858e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.420813] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.421085] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.421911] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6eee33c-b587-4fb7-9a02-78bf9d0cb692 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.430394] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 788.430394] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a380a5-90ee-fa00-f1ca-b8606bbcefa0" [ 788.430394] env[63418]: _type = "Task" [ 788.430394] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.437932] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64062a28-1f75-4025-872c-fe3b5f9e0025 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.450577] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a380a5-90ee-fa00-f1ca-b8606bbcefa0, 'name': SearchDatastore_Task, 'duration_secs': 0.010741} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.451593] env[63418]: DEBUG nova.network.neutron [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Updating instance_info_cache with network_info: [{"id": "75354cd4-70a1-4aa8-aab4-6ae9cb843f5a", "address": "fa:16:3e:c4:40:82", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75354cd4-70", "ovs_interfaceid": "75354cd4-70a1-4aa8-aab4-6ae9cb843f5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.461968] env[63418]: DEBUG nova.compute.provider_tree [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.463453] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a3f1019-b6f0-4e27-86aa-3ede5ef3f6c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.471455] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 788.471455] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bcd403-0090-16a9-9516-70a12e7f33f9" [ 788.471455] env[63418]: _type = "Task" [ 788.471455] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.484702] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bcd403-0090-16a9-9516-70a12e7f33f9, 'name': SearchDatastore_Task, 'duration_secs': 0.00918} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.485015] env[63418]: DEBUG oslo_concurrency.lockutils [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.485297] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk. {{(pid=63418) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 788.485571] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad3036e6-842e-43a8-ac7e-66a7a96d9f7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.493802] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 788.493802] env[63418]: value = "task-1244854" [ 788.493802] env[63418]: _type = "Task" [ 788.493802] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.502957] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.699689] env[63418]: DEBUG oslo_vmware.api [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244850, 'name': PowerOnVM_Task, 'duration_secs': 0.764206} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.700177] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.700583] env[63418]: INFO nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Took 9.75 seconds to spawn the instance on the hypervisor. [ 788.700833] env[63418]: DEBUG nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 788.701799] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d412a9a-1e93-4288-8280-6f4ac788591f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.722261] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244841, 'name': RelocateVM_Task, 'duration_secs': 4.182752} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.723151] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 788.724527] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268382', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'name': 'volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'serial': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 788.724527] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2b7a9a-45ab-4ce1-8bca-1e4811b38028 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.747044] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450de0dc-4a6c-43e0-9f54-c1d31f5a98f3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.774332] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce/volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.775641] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2fc6ee10-0135-4185-ba23-ebfb80b89c69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.801038] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 788.801038] env[63418]: value = "task-1244855" [ 788.801038] env[63418]: _type = "Task" [ 788.801038] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.813236] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.857023] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244852, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49665} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.857433] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 18327b62-d022-45e5-9ab0-71791491e61f/18327b62-d022-45e5-9ab0-71791491e61f.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.857762] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.858142] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee770b18-393a-48fb-bac1-c03048a6dd56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.866707] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 788.866707] env[63418]: value = "task-1244856" [ 788.866707] env[63418]: _type = "Task" [ 788.866707] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.878369] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.964431] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "refresh_cache-0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.965118] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Instance network_info: |[{"id": "75354cd4-70a1-4aa8-aab4-6ae9cb843f5a", "address": "fa:16:3e:c4:40:82", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75354cd4-70", "ovs_interfaceid": "75354cd4-70a1-4aa8-aab4-6ae9cb843f5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 788.965512] env[63418]: DEBUG oslo_concurrency.lockutils [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] Acquired lock "refresh_cache-0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.965889] env[63418]: DEBUG nova.network.neutron [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Refreshing network info cache for port 75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.967732] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:40:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75354cd4-70a1-4aa8-aab4-6ae9cb843f5a', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.979906] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating folder: Project (d445600834dd4c7e8022349ee993f3ef). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.984353] env[63418]: DEBUG nova.scheduler.client.report [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 788.988743] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 788.991056] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74380c62-9154-4ced-82bc-bdc86479ec12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.006429] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244854, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.011217] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created folder: Project (d445600834dd4c7e8022349ee993f3ef) in parent group-v268354. [ 789.011530] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating folder: Instances. Parent ref: group-v268419. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 789.013968] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f21f5a6-92a9-4e1f-81c5-6c7aa78c40f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.023068] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.023378] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.023541] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.023721] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.023901] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.024073] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.024289] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.024447] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.024616] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.024794] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.024957] env[63418]: DEBUG nova.virt.hardware [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.026317] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366b1f12-88ee-4443-96ad-41b55b7d5e22 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.032928] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created folder: Instances in parent group-v268419. [ 789.033262] env[63418]: DEBUG oslo.service.loopingcall [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.034212] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 789.034442] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50c3eaf0-0dd1-4d41-a6cc-15f32cf80f56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.053285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c14ab33-e289-4db7-8f9c-75cadd982d17 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.060849] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.060849] env[63418]: value = "task-1244859" [ 789.060849] env[63418]: _type = "Task" [ 789.060849] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.083424] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244859, 'name': CreateVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.230828] env[63418]: INFO nova.compute.manager [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Took 37.24 seconds to build instance. [ 789.248262] env[63418]: DEBUG nova.network.neutron [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Updated VIF entry in instance network info cache for port 75354cd4-70a1-4aa8-aab4-6ae9cb843f5a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 789.249396] env[63418]: DEBUG nova.network.neutron [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Updating instance_info_cache with network_info: [{"id": "75354cd4-70a1-4aa8-aab4-6ae9cb843f5a", "address": "fa:16:3e:c4:40:82", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75354cd4-70", "ovs_interfaceid": "75354cd4-70a1-4aa8-aab4-6ae9cb843f5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.311825] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244855, 'name': ReconfigVM_Task, 'duration_secs': 0.457083} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.315020] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Reconfigured VM instance instance-00000039 to attach disk [datastore1] volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce/volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.317309] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29750c58-770f-4a16-bae5-9ff6c03d4bd6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.342105] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 789.342105] env[63418]: value = "task-1244860" [ 789.342105] env[63418]: _type = "Task" [ 789.342105] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.354385] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.378015] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077403} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.378462] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.379335] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d37db1-c717-45a9-bb9e-4ceb41240f04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.403532] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 18327b62-d022-45e5-9ab0-71791491e61f/18327b62-d022-45e5-9ab0-71791491e61f.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.405259] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-140e7ca5-ed7b-4440-9fd2-5c38142a2f94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.424572] env[63418]: DEBUG nova.compute.manager [req-56ae0970-7550-4134-9e76-06582b31e0c3 req-932ebeb7-db4a-49e6-9109-ce4882eda843 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Received event network-vif-plugged-f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 789.424572] env[63418]: DEBUG oslo_concurrency.lockutils [req-56ae0970-7550-4134-9e76-06582b31e0c3 req-932ebeb7-db4a-49e6-9109-ce4882eda843 service nova] Acquiring lock "27290558-6c58-414d-bd53-c4a686890721-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.424572] env[63418]: DEBUG oslo_concurrency.lockutils [req-56ae0970-7550-4134-9e76-06582b31e0c3 req-932ebeb7-db4a-49e6-9109-ce4882eda843 service nova] Lock "27290558-6c58-414d-bd53-c4a686890721-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.424572] env[63418]: DEBUG oslo_concurrency.lockutils [req-56ae0970-7550-4134-9e76-06582b31e0c3 req-932ebeb7-db4a-49e6-9109-ce4882eda843 service nova] Lock "27290558-6c58-414d-bd53-c4a686890721-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.424572] env[63418]: DEBUG nova.compute.manager [req-56ae0970-7550-4134-9e76-06582b31e0c3 req-932ebeb7-db4a-49e6-9109-ce4882eda843 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] No waiting events found dispatching network-vif-plugged-f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 789.425040] env[63418]: WARNING nova.compute.manager [req-56ae0970-7550-4134-9e76-06582b31e0c3 req-932ebeb7-db4a-49e6-9109-ce4882eda843 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Received unexpected event network-vif-plugged-f161fb5a-0ae6-4123-8922-c602842c4933 for instance with vm_state building and task_state spawning. [ 789.430450] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 789.430450] env[63418]: value = "task-1244861" [ 789.430450] env[63418]: _type = "Task" [ 789.430450] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.442238] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244861, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.494865] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.495396] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 789.498467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.122s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.500007] env[63418]: INFO nova.compute.claims [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.511387] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.511676] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk. [ 789.512533] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b58def-e0a3-4c6f-86ca-4f13bff8e2a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.539252] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.540725] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af92cb41-afaf-453d-b0a2-6ee243050019 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.561511] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 789.561511] env[63418]: value = "task-1244862" [ 789.561511] env[63418]: _type = "Task" [ 789.561511] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.574135] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244862, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.583464] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244859, 'name': CreateVM_Task, 'duration_secs': 0.408367} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.583627] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.584354] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.584510] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.584944] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.585121] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f7ddf3d-5da5-4ef6-891e-935bb1d10ac4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.591372] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 789.591372] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526d9912-56bf-84c7-dcbf-6e826bacf34e" [ 789.591372] env[63418]: _type = "Task" [ 789.591372] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.600181] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526d9912-56bf-84c7-dcbf-6e826bacf34e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.733429] env[63418]: DEBUG oslo_concurrency.lockutils [None req-631a5c66-8eb6-4cd1-b5af-555d400415c5 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.748s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.752858] env[63418]: DEBUG oslo_concurrency.lockutils [req-fb356863-cf80-446e-abc1-07aa0fa2af20 req-98def404-a235-454a-bd30-9571409e3255 service nova] Releasing lock "refresh_cache-0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.854044] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244860, 'name': ReconfigVM_Task, 'duration_secs': 0.253082} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.854487] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268382', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'name': 'volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'serial': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 789.855100] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-937949dc-d89e-498d-8723-3380e222195f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.857407] env[63418]: DEBUG nova.network.neutron [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Successfully updated port: f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 789.865823] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 789.865823] env[63418]: value = "task-1244863" [ 789.865823] env[63418]: _type = "Task" [ 789.865823] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.875942] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244863, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.941477] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244861, 'name': ReconfigVM_Task, 'duration_secs': 0.295023} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.941857] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 18327b62-d022-45e5-9ab0-71791491e61f/18327b62-d022-45e5-9ab0-71791491e61f.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.942478] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-052d3f41-f8ce-4077-8172-53881f55a019 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.950190] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 789.950190] env[63418]: value = "task-1244864" [ 789.950190] env[63418]: _type = "Task" [ 789.950190] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.959168] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244864, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.005310] env[63418]: DEBUG nova.compute.utils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.006894] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 790.007379] env[63418]: DEBUG nova.network.neutron [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.047522] env[63418]: DEBUG nova.compute.manager [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Received event network-changed-f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 790.047767] env[63418]: DEBUG nova.compute.manager [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Refreshing instance network info cache due to event network-changed-f161fb5a-0ae6-4123-8922-c602842c4933. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 790.047965] env[63418]: DEBUG oslo_concurrency.lockutils [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] Acquiring lock "refresh_cache-27290558-6c58-414d-bd53-c4a686890721" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.048068] env[63418]: DEBUG oslo_concurrency.lockutils [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] Acquired lock "refresh_cache-27290558-6c58-414d-bd53-c4a686890721" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.048255] env[63418]: DEBUG nova.network.neutron [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Refreshing network info cache for port f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.058141] env[63418]: DEBUG nova.policy [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '688cbb2075654247959d98a2eec2b2d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '334e6dbe6b0e4ba1917396fb73cf58c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.071692] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244862, 'name': ReconfigVM_Task, 'duration_secs': 0.317899} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.072016] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Reconfigured VM instance instance-00000037 to attach disk [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.072879] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e22b2e-c9ad-4100-b55a-867341e9ac07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.106544] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce97cfc1-5eff-4c77-a654-1780256ae62a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.123742] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526d9912-56bf-84c7-dcbf-6e826bacf34e, 'name': SearchDatastore_Task, 'duration_secs': 0.011114} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.125287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.125527] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.125761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.125909] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.126452] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.126539] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 790.126539] env[63418]: value = "task-1244865" [ 790.126539] env[63418]: _type = "Task" [ 790.126539] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.127191] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c379cee1-bc7c-4707-bfe4-c537401b6d76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.143129] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244865, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.144320] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.144428] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.145216] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-501ea965-78e9-43f3-bcdc-818c7baf4100 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.151751] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 790.151751] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5245536b-66b7-a3cf-0a71-dae8ad869b7a" [ 790.151751] env[63418]: _type = "Task" [ 790.151751] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.160965] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5245536b-66b7-a3cf-0a71-dae8ad869b7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.239021] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 790.360455] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "refresh_cache-27290558-6c58-414d-bd53-c4a686890721" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.361644] env[63418]: DEBUG nova.network.neutron [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Successfully created port: 9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.375580] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244863, 'name': Rename_Task, 'duration_secs': 0.430819} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.375842] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 790.376157] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5849ce30-68ca-4838-ac97-28455437ab99 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.384067] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 790.384067] env[63418]: value = "task-1244866" [ 790.384067] env[63418]: _type = "Task" [ 790.384067] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.392789] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244866, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.462706] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244864, 'name': Rename_Task, 'duration_secs': 0.141727} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.463074] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 790.463358] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34b2c65f-c0ec-4f17-95da-903d00fda4df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.471599] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 790.471599] env[63418]: value = "task-1244867" [ 790.471599] env[63418]: _type = "Task" [ 790.471599] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.480351] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.513222] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 790.613509] env[63418]: DEBUG nova.network.neutron [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.640461] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244865, 'name': ReconfigVM_Task, 'duration_secs': 0.172307} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.644096] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 790.645437] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd238bfb-c173-483f-aeff-9b897c28bd32 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.653294] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 790.653294] env[63418]: value = "task-1244868" [ 790.653294] env[63418]: _type = "Task" [ 790.653294] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.671246] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.671521] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5245536b-66b7-a3cf-0a71-dae8ad869b7a, 'name': SearchDatastore_Task, 'duration_secs': 0.011683} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.677777] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6bb1952-0b39-444d-9faf-49a616a15ac8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.687322] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 790.687322] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52807166-bcd6-0d90-5165-a6eda65f269e" [ 790.687322] env[63418]: _type = "Task" [ 790.687322] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.696455] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52807166-bcd6-0d90-5165-a6eda65f269e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.751110] env[63418]: DEBUG nova.network.neutron [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.760251] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.802035] env[63418]: INFO nova.compute.manager [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Rebuilding instance [ 790.850479] env[63418]: DEBUG nova.compute.manager [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 790.851476] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a31888-5cbb-446f-85ae-56e4e5304ae7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.886224] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b35bb5-98ba-46b6-9ed8-2486a1bf0c24 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.901543] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478348f1-06da-43bd-86be-e53b8207d29c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.905472] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244866, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.935036] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8151391f-d958-4f1f-9b6e-e0a954a5b867 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.945234] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c759b3ac-3725-4f07-b70e-43ee0cd8ea69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.966262] env[63418]: DEBUG nova.compute.provider_tree [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.986405] env[63418]: DEBUG oslo_vmware.api [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244867, 'name': PowerOnVM_Task, 'duration_secs': 0.505164} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.986405] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.986405] env[63418]: INFO nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Took 7.13 seconds to spawn the instance on the hypervisor. [ 790.986405] env[63418]: DEBUG nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 790.987270] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a099513-31d0-4fc3-ae1d-b30ae09c65e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.167823] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244868, 'name': PowerOnVM_Task} progress is 74%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.198843] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52807166-bcd6-0d90-5165-a6eda65f269e, 'name': SearchDatastore_Task, 'duration_secs': 0.010348} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.199127] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.199386] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48/0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.199643] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-594526f1-2f65-46b4-a1a9-bd5683fd477c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.208318] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 791.208318] env[63418]: value = "task-1244869" [ 791.208318] env[63418]: _type = "Task" [ 791.208318] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.217938] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.254076] env[63418]: DEBUG oslo_concurrency.lockutils [req-94864cc7-45d2-468e-bf76-ffedf7df23cc req-93f880af-33a4-42b4-b6fe-e8403be0dba5 service nova] Releasing lock "refresh_cache-27290558-6c58-414d-bd53-c4a686890721" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.254490] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquired lock "refresh_cache-27290558-6c58-414d-bd53-c4a686890721" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.254658] env[63418]: DEBUG nova.network.neutron [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.396066] env[63418]: DEBUG oslo_vmware.api [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1244866, 'name': PowerOnVM_Task, 'duration_secs': 0.649792} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.396361] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 791.396656] env[63418]: INFO nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Took 8.27 seconds to spawn the instance on the hypervisor. [ 791.396805] env[63418]: DEBUG nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 791.397659] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987d76e2-6daa-4c2c-adaa-bb1443ff9a57 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.470641] env[63418]: DEBUG nova.scheduler.client.report [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 791.507462] env[63418]: INFO nova.compute.manager [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Took 35.71 seconds to build instance. [ 791.529132] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 791.558154] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.558556] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.558762] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.559009] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.559225] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.559403] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.559653] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.559855] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.560071] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.560277] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.560492] env[63418]: DEBUG nova.virt.hardware [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.561435] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3641a535-ce66-4c2a-b92a-ae585749f797 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.572035] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6e01e2-4f7f-4c51-89a4-e8b4cc081339 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.668093] env[63418]: DEBUG oslo_vmware.api [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244868, 'name': PowerOnVM_Task, 'duration_secs': 0.625818} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.668384] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 791.671076] env[63418]: DEBUG nova.compute.manager [None req-59d40caf-a0c7-464e-9556-f60fa90762c6 tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 791.671950] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b39610a-bcd2-4b73-8f3b-81fa5fca970e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.721464] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445625} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.721776] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48/0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.722040] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.722329] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9529789a-ed96-4ca6-8b60-a939721445d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.730573] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 791.730573] env[63418]: value = "task-1244870" [ 791.730573] env[63418]: _type = "Task" [ 791.730573] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.739847] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.792557] env[63418]: DEBUG nova.network.neutron [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.870092] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 791.870423] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34af93c7-f311-4697-8372-6ab36a0af074 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.881620] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 791.881620] env[63418]: value = "task-1244871" [ 791.881620] env[63418]: _type = "Task" [ 791.881620] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.893847] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244871, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.918286] env[63418]: INFO nova.compute.manager [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Took 38.07 seconds to build instance. [ 791.943852] env[63418]: DEBUG nova.network.neutron [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Successfully updated port: 9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.976573] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.978299] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 791.980498] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.227s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.981114] env[63418]: DEBUG nova.objects.instance [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lazy-loading 'resources' on Instance uuid c0fd1999-346d-4c12-9b8c-ab7e21ec4227 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.009715] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25816ba0-8296-4016-b255-2d8c0eb0c88f tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "18327b62-d022-45e5-9ab0-71791491e61f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.573s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.063570] env[63418]: DEBUG nova.network.neutron [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Updating instance_info_cache with network_info: [{"id": "f161fb5a-0ae6-4123-8922-c602842c4933", "address": "fa:16:3e:d1:b0:c6", "network": {"id": "8785e223-f98b-44b2-837d-fbaa4e5b546f", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1906109871-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88d4c2f44b5641acb3ebbf83967e1beb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91712705-510f-41a0-a803-2ecd92b676e1", "external-id": "nsx-vlan-transportzone-512", "segmentation_id": 512, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf161fb5a-0a", "ovs_interfaceid": "f161fb5a-0ae6-4123-8922-c602842c4933", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.201894] env[63418]: DEBUG nova.compute.manager [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Received event network-vif-plugged-9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 792.202156] env[63418]: DEBUG oslo_concurrency.lockutils [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] Acquiring lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.202505] env[63418]: DEBUG oslo_concurrency.lockutils [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.202573] env[63418]: DEBUG oslo_concurrency.lockutils [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.202731] env[63418]: DEBUG nova.compute.manager [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] No waiting events found dispatching network-vif-plugged-9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 792.202892] env[63418]: WARNING nova.compute.manager [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Received unexpected event network-vif-plugged-9b941608-e1b2-4bce-8249-e53c075ac132 for instance with vm_state building and task_state spawning. [ 792.203492] env[63418]: DEBUG nova.compute.manager [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Received event network-changed-9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 792.203693] env[63418]: DEBUG nova.compute.manager [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Refreshing instance network info cache due to event network-changed-9b941608-e1b2-4bce-8249-e53c075ac132. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 792.203939] env[63418]: DEBUG oslo_concurrency.lockutils [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] Acquiring lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.204101] env[63418]: DEBUG oslo_concurrency.lockutils [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] Acquired lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.204286] env[63418]: DEBUG nova.network.neutron [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Refreshing network info cache for port 9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.241621] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.242034] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.243421] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a435e5b8-0cee-49fe-8a8c-c06312416642 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.272822] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48/0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.273240] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2d596cf-3699-4879-90b5-9583e4d707fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.297427] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 792.297427] env[63418]: value = "task-1244872" [ 792.297427] env[63418]: _type = "Task" [ 792.297427] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.306569] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244872, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.325121] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "18327b62-d022-45e5-9ab0-71791491e61f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.325363] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "18327b62-d022-45e5-9ab0-71791491e61f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.325606] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "18327b62-d022-45e5-9ab0-71791491e61f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.325816] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "18327b62-d022-45e5-9ab0-71791491e61f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.326124] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "18327b62-d022-45e5-9ab0-71791491e61f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.328570] env[63418]: INFO nova.compute.manager [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Terminating instance [ 792.392388] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244871, 'name': PowerOffVM_Task, 'duration_secs': 0.463994} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.392708] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 792.393133] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.394434] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1948f58c-d319-4b44-a80d-f0936174bf97 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.402627] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 792.402627] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-466689a3-d04d-47f6-a649-d8c433a1e0c1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.420505] env[63418]: DEBUG oslo_concurrency.lockutils [None req-de9c65e2-ba41-45e6-89e2-79ffe4817373 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.523s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.446494] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.484254] env[63418]: DEBUG nova.compute.utils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.491231] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 792.491604] env[63418]: DEBUG nova.network.neutron [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.511880] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 792.566512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Releasing lock "refresh_cache-27290558-6c58-414d-bd53-c4a686890721" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.567031] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Instance network_info: |[{"id": "f161fb5a-0ae6-4123-8922-c602842c4933", "address": "fa:16:3e:d1:b0:c6", "network": {"id": "8785e223-f98b-44b2-837d-fbaa4e5b546f", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1906109871-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88d4c2f44b5641acb3ebbf83967e1beb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91712705-510f-41a0-a803-2ecd92b676e1", "external-id": "nsx-vlan-transportzone-512", "segmentation_id": 512, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf161fb5a-0a", "ovs_interfaceid": "f161fb5a-0ae6-4123-8922-c602842c4933", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 792.567580] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:b0:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91712705-510f-41a0-a803-2ecd92b676e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f161fb5a-0ae6-4123-8922-c602842c4933', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.575986] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Creating folder: Project (88d4c2f44b5641acb3ebbf83967e1beb). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.578906] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8898dfc-2b29-47a4-ac11-f89a3d643ab5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.582167] env[63418]: DEBUG nova.policy [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.594755] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Created folder: Project (88d4c2f44b5641acb3ebbf83967e1beb) in parent group-v268354. [ 792.595014] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Creating folder: Instances. Parent ref: group-v268422. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.595296] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aeeb2fc6-826e-48c9-a178-82b7ef1256a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.607652] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Created folder: Instances in parent group-v268422. [ 792.608104] env[63418]: DEBUG oslo.service.loopingcall [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.608427] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27290558-6c58-414d-bd53-c4a686890721] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.608772] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1e5ed3d-1573-4ce2-b812-32891e01602d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.643680] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.643680] env[63418]: value = "task-1244876" [ 792.643680] env[63418]: _type = "Task" [ 792.643680] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.658125] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244876, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.737038] env[63418]: DEBUG nova.network.neutron [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.809800] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.817888] env[63418]: DEBUG nova.network.neutron [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.832268] env[63418]: DEBUG nova.compute.manager [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 792.832532] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.833486] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0875bc3-7a1b-4576-a770-8904c57ef55f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.842569] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 792.842809] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57191c69-0e01-4dbd-9598-0130237e4543 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.850358] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 792.850358] env[63418]: value = "task-1244877" [ 792.850358] env[63418]: _type = "Task" [ 792.850358] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.856285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2966e248-5154-4493-a61a-a283fe4927c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.865159] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.868290] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f568f0e7-9e0e-4071-adc8-ddcc99749d46 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.901600] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a431064b-6eaa-41de-9e80-b131de91206f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.909828] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d474156-13f1-435f-8f82-f06c8f23af7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.928229] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 792.929763] env[63418]: DEBUG nova.compute.provider_tree [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.991477] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 793.033339] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.156929] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244876, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.273731] env[63418]: DEBUG nova.network.neutron [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Successfully created port: 1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.312729] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244872, 'name': ReconfigVM_Task, 'duration_secs': 1.008584} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.313071] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48/0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.313792] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-727d82b6-e227-4869-b311-27cbe938ede8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.320531] env[63418]: DEBUG oslo_concurrency.lockutils [req-cf855d97-961f-4625-adc0-f1960b8de9f0 req-22731890-932f-4e39-8fba-6a738468832e service nova] Releasing lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.321869] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquired lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.322780] env[63418]: DEBUG nova.network.neutron [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.327022] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 793.327022] env[63418]: value = "task-1244878" [ 793.327022] env[63418]: _type = "Task" [ 793.327022] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.333716] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 793.333963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 793.334473] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore1] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.335361] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d341ea4c-b0b9-4e71-b3a1-3888999376d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.343696] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244878, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.350420] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 793.350420] env[63418]: value = "task-1244879" [ 793.350420] env[63418]: _type = "Task" [ 793.350420] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.364745] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244879, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.368553] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244877, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.435846] env[63418]: DEBUG nova.scheduler.client.report [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 793.467820] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.582964] env[63418]: DEBUG nova.compute.manager [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 793.583369] env[63418]: DEBUG nova.compute.manager [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing instance network info cache due to event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 793.583914] env[63418]: DEBUG oslo_concurrency.lockutils [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.584250] env[63418]: DEBUG oslo_concurrency.lockutils [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.584464] env[63418]: DEBUG nova.network.neutron [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 793.655953] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244876, 'name': CreateVM_Task, 'duration_secs': 0.94257} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.655953] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27290558-6c58-414d-bd53-c4a686890721] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.656567] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.656756] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.657129] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.657418] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dda46aa2-eac9-40d8-9405-c45b2cebc212 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.663103] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 793.663103] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529846a0-aae2-e95f-ab68-923d6f9ae164" [ 793.663103] env[63418]: _type = "Task" [ 793.663103] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.672623] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529846a0-aae2-e95f-ab68-923d6f9ae164, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.843026] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244878, 'name': Rename_Task, 'duration_secs': 0.193663} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.843026] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 793.843026] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab510b8e-e4bd-4bc0-9970-46ca51315115 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.851841] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 793.851841] env[63418]: value = "task-1244880" [ 793.851841] env[63418]: _type = "Task" [ 793.851841] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.870786] env[63418]: DEBUG nova.network.neutron [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.881810] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244879, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141533} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.890471] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.891047] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 793.891491] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.898162] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244877, 'name': PowerOffVM_Task, 'duration_secs': 0.562288} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.898162] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244880, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.898162] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 793.901033] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 793.901132] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bbfffac-0479-4864-b862-a51858398474 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.948618] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.968s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.951089] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.435s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.953341] env[63418]: INFO nova.compute.claims [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.985027] env[63418]: INFO nova.scheduler.client.report [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleted allocations for instance c0fd1999-346d-4c12-9b8c-ab7e21ec4227 [ 793.994938] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 793.995543] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 793.995543] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Deleting the datastore file [datastore1] 18327b62-d022-45e5-9ab0-71791491e61f {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.996103] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6300351-a072-424e-93d5-bd3a2aeb1d86 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.006093] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for the task: (returnval){ [ 794.006093] env[63418]: value = "task-1244882" [ 794.006093] env[63418]: _type = "Task" [ 794.006093] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.007928] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 794.025577] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.046872] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.047061] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.047105] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.047288] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.047447] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.047595] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.047823] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.048011] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.048219] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.048541] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.048628] env[63418]: DEBUG nova.virt.hardware [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.049579] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0a1a70-0f37-48da-b75a-fb42491d3635 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.054478] env[63418]: DEBUG nova.network.neutron [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Updating instance_info_cache with network_info: [{"id": "9b941608-e1b2-4bce-8249-e53c075ac132", "address": "fa:16:3e:78:59:7e", "network": {"id": "3cda032d-334b-4784-88a8-5f0dcc78e7ea", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1114152170-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "334e6dbe6b0e4ba1917396fb73cf58c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b941608-e1", "ovs_interfaceid": "9b941608-e1b2-4bce-8249-e53c075ac132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.063847] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41df29b-f1ce-403f-ad7d-f796070176d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.174708] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529846a0-aae2-e95f-ab68-923d6f9ae164, 'name': SearchDatastore_Task, 'duration_secs': 0.009468} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.175051] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.175294] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.175523] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.175789] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.175850] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.176109] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f21bff43-e62f-40ca-b230-7c89d57800da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.186159] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.186395] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 794.187322] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a99a6a75-d0e7-466c-8c4f-7157b681e692 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.196211] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 794.196211] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529c10e0-8b8a-3391-a972-c886ef746201" [ 794.196211] env[63418]: _type = "Task" [ 794.196211] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.205079] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529c10e0-8b8a-3391-a972-c886ef746201, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.313577] env[63418]: DEBUG nova.network.neutron [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updated VIF entry in instance network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 794.313966] env[63418]: DEBUG nova.network.neutron [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.376680] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244880, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.494914] env[63418]: DEBUG oslo_concurrency.lockutils [None req-182ae20f-412e-4f64-bbbd-570e93dbe1eb tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "c0fd1999-346d-4c12-9b8c-ab7e21ec4227" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.221s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.524666] env[63418]: DEBUG oslo_vmware.api [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Task: {'id': task-1244882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136507} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.525015] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.525252] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.525463] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.525666] env[63418]: INFO nova.compute.manager [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Took 1.69 seconds to destroy the instance on the hypervisor. [ 794.525942] env[63418]: DEBUG oslo.service.loopingcall [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.526193] env[63418]: DEBUG nova.compute.manager [-] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 794.526336] env[63418]: DEBUG nova.network.neutron [-] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.557347] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Releasing lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.557765] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Instance network_info: |[{"id": "9b941608-e1b2-4bce-8249-e53c075ac132", "address": "fa:16:3e:78:59:7e", "network": {"id": "3cda032d-334b-4784-88a8-5f0dcc78e7ea", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1114152170-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "334e6dbe6b0e4ba1917396fb73cf58c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b941608-e1", "ovs_interfaceid": "9b941608-e1b2-4bce-8249-e53c075ac132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 794.558165] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:59:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8233b50c-be21-441a-a357-a29138a95b8b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b941608-e1b2-4bce-8249-e53c075ac132', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.566522] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Creating folder: Project (334e6dbe6b0e4ba1917396fb73cf58c8). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.566943] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1dfbe431-e5eb-452f-a181-0d23afc7055a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.582174] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Created folder: Project (334e6dbe6b0e4ba1917396fb73cf58c8) in parent group-v268354. [ 794.582401] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Creating folder: Instances. Parent ref: group-v268425. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.582749] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c74632e-bc72-4ae3-b703-fbe2c9b74a31 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.594277] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Created folder: Instances in parent group-v268425. [ 794.594277] env[63418]: DEBUG oslo.service.loopingcall [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.594429] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.594820] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df4d9cef-8af1-4da9-ab5b-78e1dbeb3565 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.616317] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.616317] env[63418]: value = "task-1244885" [ 794.616317] env[63418]: _type = "Task" [ 794.616317] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.625890] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244885, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.707215] env[63418]: DEBUG nova.compute.manager [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Received event network-changed-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 794.707427] env[63418]: DEBUG nova.compute.manager [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Refreshing instance network info cache due to event network-changed-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 794.707932] env[63418]: DEBUG oslo_concurrency.lockutils [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] Acquiring lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.707932] env[63418]: DEBUG oslo_concurrency.lockutils [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] Acquired lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.708197] env[63418]: DEBUG nova.network.neutron [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Refreshing network info cache for port 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.715814] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529c10e0-8b8a-3391-a972-c886ef746201, 'name': SearchDatastore_Task, 'duration_secs': 0.009747} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.716696] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b2a37f7-163a-4e7f-8003-6721a73d5e8d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.725218] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 794.725218] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5205fe10-4826-d90c-063b-47255b07c7d1" [ 794.725218] env[63418]: _type = "Task" [ 794.725218] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.734152] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5205fe10-4826-d90c-063b-47255b07c7d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.816386] env[63418]: DEBUG oslo_concurrency.lockutils [req-0f349888-136d-45c6-9826-9f494d482c37 req-85af7f85-ab1a-443e-9601-a6ba38847782 service nova] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.878139] env[63418]: DEBUG oslo_vmware.api [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244880, 'name': PowerOnVM_Task, 'duration_secs': 0.754885} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.878471] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.878664] env[63418]: INFO nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Took 8.59 seconds to spawn the instance on the hypervisor. [ 794.878843] env[63418]: DEBUG nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 794.879658] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e41b5f-2815-48ce-a813-c32c7ec185b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.931840] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.931840] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.931840] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.931998] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.931998] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.931998] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.932104] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.932240] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.932407] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.932567] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.932735] env[63418]: DEBUG nova.virt.hardware [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.933657] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52b6c46-ebf5-4967-a830-f2c78c05f2ad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.943596] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e2b002-80d7-4d72-b934-1006a8d7dcbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.958014] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:99:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7766654-bf8e-4a36-b935-da5442b28a70', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.965510] env[63418]: DEBUG oslo.service.loopingcall [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.968251] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.969638] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe657492-102b-4b33-a206-6e049755a6a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.991355] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.991355] env[63418]: value = "task-1244886" [ 794.991355] env[63418]: _type = "Task" [ 794.991355] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.002196] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244886, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.128633] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244885, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.241196] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5205fe10-4826-d90c-063b-47255b07c7d1, 'name': SearchDatastore_Task, 'duration_secs': 0.01152} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.241196] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.241196] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 27290558-6c58-414d-bd53-c4a686890721/27290558-6c58-414d-bd53-c4a686890721.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 795.241196] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edbe8dcc-e7dc-40b0-95c0-97cf52545d13 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.254332] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 795.254332] env[63418]: value = "task-1244887" [ 795.254332] env[63418]: _type = "Task" [ 795.254332] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.269183] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.284806] env[63418]: DEBUG nova.network.neutron [-] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.315527] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8913dbf6-4432-4078-80f6-3daa6bada0f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.323837] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4e3fbb-7c4f-4dca-b343-10c261a281ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.360139] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566a34c3-d21c-46fc-96dc-c4e75994b0fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.368872] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b8081e6-64a0-4f1d-9496-94daca869186 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.384779] env[63418]: DEBUG nova.compute.provider_tree [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.398761] env[63418]: INFO nova.compute.manager [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Took 38.69 seconds to build instance. [ 795.504073] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244886, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.534276] env[63418]: DEBUG nova.network.neutron [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updated VIF entry in instance network info cache for port 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 795.534670] env[63418]: DEBUG nova.network.neutron [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updating instance_info_cache with network_info: [{"id": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "address": "fa:16:3e:fc:88:79", "network": {"id": "4d39fd83-617d-47b7-a1e6-603411a20c2b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1579674115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02f73883249c44368eaaca666eacffc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72cfead0-4e", "ovs_interfaceid": "72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.543554] env[63418]: DEBUG nova.network.neutron [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Successfully updated port: 1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.629079] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244885, 'name': CreateVM_Task, 'duration_secs': 0.607716} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.629572] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 795.630402] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.630580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.630986] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.631225] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a72da70-c098-45b6-9973-94e00753cc29 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.637806] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 795.637806] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523006b8-2da2-d0e5-e6da-f22cc921222a" [ 795.637806] env[63418]: _type = "Task" [ 795.637806] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.647808] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523006b8-2da2-d0e5-e6da-f22cc921222a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.768834] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482119} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.769668] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 27290558-6c58-414d-bd53-c4a686890721/27290558-6c58-414d-bd53-c4a686890721.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.769668] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.769668] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5e21d99-e001-4782-b7f8-d732659ebbb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.778331] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 795.778331] env[63418]: value = "task-1244888" [ 795.778331] env[63418]: _type = "Task" [ 795.778331] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.789220] env[63418]: INFO nova.compute.manager [-] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Took 1.26 seconds to deallocate network for instance. [ 795.789482] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244888, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.891156] env[63418]: DEBUG nova.scheduler.client.report [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 795.902605] env[63418]: DEBUG oslo_concurrency.lockutils [None req-479cef16-fb3b-4e2d-8744-5f32585e9aa4 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.307s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.003444] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244886, 'name': CreateVM_Task, 'duration_secs': 0.527172} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.003669] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 796.004401] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.037665] env[63418]: DEBUG oslo_concurrency.lockutils [req-863e4784-9dbf-4ef2-a716-4fd654ffc64c req-c5635ae9-4848-4b9d-8f4a-fa1faa9ce056 service nova] Releasing lock "refresh_cache-1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.044843] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.045052] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.045341] env[63418]: DEBUG nova.network.neutron [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.094732] env[63418]: DEBUG nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 796.094934] env[63418]: DEBUG nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing instance network info cache due to event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 796.095370] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.095602] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.096044] env[63418]: DEBUG nova.network.neutron [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.149542] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523006b8-2da2-d0e5-e6da-f22cc921222a, 'name': SearchDatastore_Task, 'duration_secs': 0.052907} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.149926] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.150198] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.150660] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.150909] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.151250] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.151390] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.151714] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.151996] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e05dd05a-e013-49fd-9ada-5181569a2339 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.154051] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6007265f-ba9b-41a9-bc5c-71a4fa973180 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.159993] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 796.159993] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5250670f-443c-41be-ea80-8731b7e9f68f" [ 796.159993] env[63418]: _type = "Task" [ 796.159993] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.164472] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.164689] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.165672] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f2a4a98-dd3c-48d6-8a4a-73d700fd41f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.172369] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5250670f-443c-41be-ea80-8731b7e9f68f, 'name': SearchDatastore_Task, 'duration_secs': 0.009682} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.172964] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.173276] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.173551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.175476] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 796.175476] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5237b223-45ac-d6bd-0e35-dfad9fd1caea" [ 796.175476] env[63418]: _type = "Task" [ 796.175476] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.182940] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5237b223-45ac-d6bd-0e35-dfad9fd1caea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.264083] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.264363] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.264545] env[63418]: INFO nova.compute.manager [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Shelving [ 796.288884] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244888, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064711} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.289341] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.290143] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435e7799-ce17-4b1d-bd98-f1cd92e91306 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.295921] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.315616] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 27290558-6c58-414d-bd53-c4a686890721/27290558-6c58-414d-bd53-c4a686890721.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.316317] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96b191cd-8e8d-4903-b75f-bf13b08726ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.337016] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 796.337016] env[63418]: value = "task-1244889" [ 796.337016] env[63418]: _type = "Task" [ 796.337016] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.345205] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.396388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.396757] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 796.404515] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.079s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.404745] env[63418]: DEBUG nova.objects.instance [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'resources' on Instance uuid c4cb1770-b3c7-4f6b-ba11-85310f2175a0 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.570412] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "17a055e4-13da-4984-ae64-af1ae9d652f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.570680] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.570892] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "17a055e4-13da-4984-ae64-af1ae9d652f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.571490] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.571490] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.573674] env[63418]: INFO nova.compute.manager [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Terminating instance [ 796.589893] env[63418]: DEBUG nova.network.neutron [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.692593] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5237b223-45ac-d6bd-0e35-dfad9fd1caea, 'name': SearchDatastore_Task, 'duration_secs': 0.008491} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.694570] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ca645f4-2056-44f3-a1b3-6601f14c658f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.706091] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 796.706091] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d59ad9-214b-e7ed-3c85-8b6b87d3716b" [ 796.706091] env[63418]: _type = "Task" [ 796.706091] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.717733] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d59ad9-214b-e7ed-3c85-8b6b87d3716b, 'name': SearchDatastore_Task, 'duration_secs': 0.010859} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.720210] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.720509] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00/7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.720805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.720990] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.721255] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebca05aa-22ef-45a9-98a6-4f89da060e95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.725636] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54d456e6-aed2-4d16-aa78-b2bb86bf478f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.737441] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.737655] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.739801] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-542b1270-cbba-4c1a-b747-55f73b83c08b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.742982] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 796.742982] env[63418]: value = "task-1244890" [ 796.742982] env[63418]: _type = "Task" [ 796.742982] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.752261] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 796.752261] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f7f551-f991-1715-7def-2aaf9b45cb84" [ 796.752261] env[63418]: _type = "Task" [ 796.752261] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.760798] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244890, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.768965] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f7f551-f991-1715-7def-2aaf9b45cb84, 'name': SearchDatastore_Task, 'duration_secs': 0.009905} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.776336] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3276f0ec-b3f4-496e-9d51-c65215a1c4ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.783267] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 796.783267] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52df6dc9-3144-1c80-7a7a-a8edfb6d6ee8" [ 796.783267] env[63418]: _type = "Task" [ 796.783267] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.793368] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52df6dc9-3144-1c80-7a7a-a8edfb6d6ee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.850162] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244889, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.867477] env[63418]: DEBUG nova.network.neutron [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updated VIF entry in instance network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 796.867900] env[63418]: DEBUG nova.network.neutron [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.906046] env[63418]: DEBUG nova.compute.utils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.907586] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 796.908067] env[63418]: DEBUG nova.network.neutron [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.922767] env[63418]: DEBUG nova.network.neutron [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.965102] env[63418]: DEBUG nova.policy [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2dcdc53fd7cc4555a4eecb6bf7564b55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4759c1d6199344c49063d35e422d6324', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.081963] env[63418]: DEBUG nova.compute.manager [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 797.081963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.083195] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f83429-ded1-4f78-a322-87e70823a2ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.096337] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.096337] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29d1503b-05ab-4bfa-85e9-2675054a608c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.103749] env[63418]: DEBUG oslo_vmware.api [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 797.103749] env[63418]: value = "task-1244891" [ 797.103749] env[63418]: _type = "Task" [ 797.103749] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.120156] env[63418]: DEBUG oslo_vmware.api [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244891, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.186697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.186969] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.187190] env[63418]: DEBUG nova.compute.manager [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 797.188107] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999f1441-abf2-4411-9038-f2c2b9c606d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.197713] env[63418]: DEBUG nova.compute.manager [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 797.198375] env[63418]: DEBUG nova.objects.instance [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lazy-loading 'flavor' on Instance uuid 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 797.254373] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244890, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484534} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.254699] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00/7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.254995] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.255264] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8e4db33-5e70-4f58-aa7c-54350ec4ff9b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.266214] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 797.266214] env[63418]: value = "task-1244892" [ 797.266214] env[63418]: _type = "Task" [ 797.266214] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.278319] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.282982] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.285503] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b29f43a-eeca-4e82-9acf-272b5398660f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.302330] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52df6dc9-3144-1c80-7a7a-a8edfb6d6ee8, 'name': SearchDatastore_Task, 'duration_secs': 0.010627} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.303294] env[63418]: DEBUG nova.network.neutron [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Successfully created port: 48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.307031] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.307031] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 797.307204] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 797.307204] env[63418]: value = "task-1244893" [ 797.307204] env[63418]: _type = "Task" [ 797.307204] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.309839] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbca00b1-49d2-4c52-a1bd-dc001bd6bacb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.325120] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.327195] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 797.327195] env[63418]: value = "task-1244894" [ 797.327195] env[63418]: _type = "Task" [ 797.327195] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.342021] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.353052] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244889, 'name': ReconfigVM_Task, 'duration_secs': 0.771282} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.353363] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 27290558-6c58-414d-bd53-c4a686890721/27290558-6c58-414d-bd53-c4a686890721.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.354048] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f485a48-eb72-4b5a-b4e2-b4a63e80f1c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.364147] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 797.364147] env[63418]: value = "task-1244895" [ 797.364147] env[63418]: _type = "Task" [ 797.364147] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.370825] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.371109] env[63418]: DEBUG nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-vif-plugged-1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 797.371379] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.371809] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.371809] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.372428] env[63418]: DEBUG nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] No waiting events found dispatching network-vif-plugged-1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 797.372645] env[63418]: WARNING nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received unexpected event network-vif-plugged-1eca951d-2df6-4179-aa0c-6babf13838f3 for instance with vm_state building and task_state spawning. [ 797.372823] env[63418]: DEBUG nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-changed-1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 797.372983] env[63418]: DEBUG nova.compute.manager [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing instance network info cache due to event network-changed-1eca951d-2df6-4179-aa0c-6babf13838f3. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 797.373175] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.379507] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244895, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.383740] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef568d71-286e-4488-87f8-1fb23865fefa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.394928] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4792f988-09b4-4bf5-802e-dbfced16c02e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.400039] env[63418]: DEBUG nova.compute.manager [req-540e9b4b-5f29-4e0a-844f-4898a4a8f52b req-ae2ab2ae-94f1-448a-94df-7c965818b6e7 service nova] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Received event network-vif-deleted-50971ce3-53f7-4832-8816-2252c1f95541 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 797.431164] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 797.434124] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.434486] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Instance network_info: |[{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 797.435147] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.435399] env[63418]: DEBUG nova.network.neutron [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing network info cache for port 1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 797.436705] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:49:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1eca951d-2df6-4179-aa0c-6babf13838f3', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.445288] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Creating folder: Project (4297b53faeab40dfa5de863ad4030800). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.446107] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b3e921-441f-4b2a-8851-40ec04788525 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.450326] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4991101-0b4c-462c-91b8-efa9e97bb79f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.461259] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4f65e9-4f5d-4988-8b1f-b9049cba4f2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.467458] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Created folder: Project (4297b53faeab40dfa5de863ad4030800) in parent group-v268354. [ 797.467724] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Creating folder: Instances. Parent ref: group-v268429. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.468922] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29e09f1f-9704-4cab-9154-2687b180c38e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.480129] env[63418]: DEBUG nova.compute.provider_tree [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.487019] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Created folder: Instances in parent group-v268429. [ 797.487019] env[63418]: DEBUG oslo.service.loopingcall [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.487019] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.487019] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6d38557-bfd7-4c85-a0c2-e843168890e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.512260] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.512260] env[63418]: value = "task-1244898" [ 797.512260] env[63418]: _type = "Task" [ 797.512260] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.522900] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244898, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.615720] env[63418]: DEBUG oslo_vmware.api [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244891, 'name': PowerOffVM_Task, 'duration_secs': 0.291715} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.616125] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.616306] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.616599] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bea071ed-274f-44dd-b09a-b4c80f9a001f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.706910] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.707323] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.707699] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleting the datastore file [datastore2] 17a055e4-13da-4984-ae64-af1ae9d652f6 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.709615] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78b1e926-9ff5-40d4-afd5-c127d76febbf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.720904] env[63418]: DEBUG oslo_vmware.api [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 797.720904] env[63418]: value = "task-1244900" [ 797.720904] env[63418]: _type = "Task" [ 797.720904] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.737114] env[63418]: DEBUG oslo_vmware.api [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244900, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.778618] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071154} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.778618] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.778618] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52b0a5b-49f6-4091-aec8-b9a79e09ea0d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.803574] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00/7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.804093] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e598323f-cfbc-4eb9-8394-e559da1921f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.830117] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244893, 'name': PowerOffVM_Task, 'duration_secs': 0.224315} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.834294] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.834710] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 797.834710] env[63418]: value = "task-1244901" [ 797.834710] env[63418]: _type = "Task" [ 797.834710] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.835413] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2faa3b86-eacc-4e11-b224-6cb7fae8a6f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.845037] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483777} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.845646] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.845876] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.846150] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e643c04-b4aa-4fff-b4cd-dfdcbbf48325 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.867532] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244901, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.872766] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44173118-a524-49e9-a337-9035456ff1a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.876314] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 797.876314] env[63418]: value = "task-1244902" [ 797.876314] env[63418]: _type = "Task" [ 797.876314] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.893904] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244895, 'name': Rename_Task, 'duration_secs': 0.236753} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.893904] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244902, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.893904] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.894138] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-044dda1b-a989-4373-91c5-fcd5f085d7b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.906641] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 797.906641] env[63418]: value = "task-1244903" [ 797.906641] env[63418]: _type = "Task" [ 797.906641] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.915971] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244903, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.985344] env[63418]: DEBUG nova.scheduler.client.report [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 798.026499] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244898, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.215219] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.215904] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e60193f5-c3c1-436b-9f37-1edb42ca0235 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.223316] env[63418]: DEBUG oslo_vmware.api [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 798.223316] env[63418]: value = "task-1244904" [ 798.223316] env[63418]: _type = "Task" [ 798.223316] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.237019] env[63418]: DEBUG oslo_vmware.api [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244900, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235684} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.241269] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.241269] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 798.241450] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 798.241506] env[63418]: INFO nova.compute.manager [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 798.242916] env[63418]: DEBUG oslo.service.loopingcall [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.242916] env[63418]: DEBUG oslo_vmware.api [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.242916] env[63418]: DEBUG nova.compute.manager [-] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 798.242916] env[63418]: DEBUG nova.network.neutron [-] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.348358] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244901, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.387824] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244902, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14571} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.388178] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.389302] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19f9db7-a3e9-43bd-a8be-12dcb5e816b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.395098] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 798.396090] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8817e121-a2a3-4db0-bbfb-d7f5074de318 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.424735] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.431971] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6acdf570-b4ae-438e-80fb-fc0ca2cee114 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.451548] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 798.451548] env[63418]: value = "task-1244905" [ 798.451548] env[63418]: _type = "Task" [ 798.451548] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.459952] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 798.461992] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244903, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.467276] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244905, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.467777] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 798.467777] env[63418]: value = "task-1244906" [ 798.467777] env[63418]: _type = "Task" [ 798.467777] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.478658] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244906, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.490762] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.500375] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.500840] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.504431] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.504431] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.504431] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.504431] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.504431] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.504821] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.504821] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.504821] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.504821] env[63418]: DEBUG nova.virt.hardware [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.504821] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.978s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.505775] env[63418]: INFO nova.compute.claims [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.513238] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58ea43e-13a5-4b75-aaa9-04cd2253ec85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.525548] env[63418]: INFO nova.scheduler.client.report [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocations for instance c4cb1770-b3c7-4f6b-ba11-85310f2175a0 [ 798.533053] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6c5004-cbf7-409f-8ab9-bbd6b179afe9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.544410] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244898, 'name': CreateVM_Task, 'duration_secs': 0.627991} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.545143] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.546058] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.546261] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.546603] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.547450] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5586f67-5fd9-471f-8319-c1e6d04ea1dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.563024] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 798.563024] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5213a3a7-c6c0-cd97-6ca9-67026f10ef84" [ 798.563024] env[63418]: _type = "Task" [ 798.563024] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.575867] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5213a3a7-c6c0-cd97-6ca9-67026f10ef84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.691108] env[63418]: DEBUG nova.network.neutron [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updated VIF entry in instance network info cache for port 1eca951d-2df6-4179-aa0c-6babf13838f3. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 798.691589] env[63418]: DEBUG nova.network.neutron [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.739039] env[63418]: DEBUG oslo_vmware.api [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244904, 'name': PowerOffVM_Task, 'duration_secs': 0.358531} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.739358] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.739569] env[63418]: DEBUG nova.compute.manager [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 798.740669] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce73dc1-0bbe-4f03-8f06-0ea12d820437 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.766256] env[63418]: DEBUG nova.compute.manager [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 798.766256] env[63418]: DEBUG nova.compute.manager [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing instance network info cache due to event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 798.766469] env[63418]: DEBUG oslo_concurrency.lockutils [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.766524] env[63418]: DEBUG oslo_concurrency.lockutils [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.766735] env[63418]: DEBUG nova.network.neutron [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.850342] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244901, 'name': ReconfigVM_Task, 'duration_secs': 0.835043} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.850652] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00/7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.851589] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1a4de06-5fc5-49f4-88f5-d0f4f1850c93 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.861557] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 798.861557] env[63418]: value = "task-1244907" [ 798.861557] env[63418]: _type = "Task" [ 798.861557] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.881009] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244907, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.923858] env[63418]: DEBUG oslo_vmware.api [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244903, 'name': PowerOnVM_Task, 'duration_secs': 0.987272} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.924233] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 798.924391] env[63418]: INFO nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Took 9.94 seconds to spawn the instance on the hypervisor. [ 798.924566] env[63418]: DEBUG nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 798.925495] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf6cfea-b4ac-459f-8d71-a14b6f12a6ef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.951409] env[63418]: DEBUG nova.network.neutron [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Successfully updated port: 48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.968634] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244905, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.980590] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.024051] env[63418]: DEBUG nova.network.neutron [-] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.036633] env[63418]: DEBUG oslo_concurrency.lockutils [None req-36c13175-1b52-4959-a679-8847778dcd2b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "c4cb1770-b3c7-4f6b-ba11-85310f2175a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.275s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.082036] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5213a3a7-c6c0-cd97-6ca9-67026f10ef84, 'name': SearchDatastore_Task, 'duration_secs': 0.031698} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.082036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.082036] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.082036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.082437] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.082437] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.082437] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29813fc3-b073-4085-90d2-1d1c17f80172 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.090515] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.090961] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.091865] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d151dcc7-37aa-48d5-94ba-617d00b972ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.099328] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 799.099328] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52083009-0e3f-bd5c-678a-2b4e39c1725b" [ 799.099328] env[63418]: _type = "Task" [ 799.099328] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.112438] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52083009-0e3f-bd5c-678a-2b4e39c1725b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.195288] env[63418]: DEBUG oslo_concurrency.lockutils [req-8cf7e330-81fb-4c71-9655-9816c4a3bee0 req-f9b6bb3e-1b21-4e01-9926-f3d335e92d7d service nova] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.255140] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8c2545d8-9317-43d0-9f05-15c069a851ae tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.068s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.376815] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244907, 'name': Rename_Task, 'duration_secs': 0.31213} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.377171] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.377452] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-005540d4-abbd-4a15-8646-129126e178f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.387423] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 799.387423] env[63418]: value = "task-1244908" [ 799.387423] env[63418]: _type = "Task" [ 799.387423] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.401615] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244908, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.447399] env[63418]: INFO nova.compute.manager [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Took 41.15 seconds to build instance. [ 799.461338] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "refresh_cache-be1c134c-f36e-4cc7-b4ef-8f30793fb4df" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.461761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "refresh_cache-be1c134c-f36e-4cc7-b4ef-8f30793fb4df" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.461761] env[63418]: DEBUG nova.network.neutron [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.469571] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244905, 'name': CreateSnapshot_Task, 'duration_secs': 0.814438} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.475142] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 799.476242] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec61aa2d-c055-4565-8997-4ba4782869cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.492823] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244906, 'name': ReconfigVM_Task, 'duration_secs': 0.650743} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.493478] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c/1f898bd4-0b80-48c1-9e7c-8948f4b19c0c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.496781] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b826360-c44e-407d-a74c-b204912ba8f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.502524] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 799.502524] env[63418]: value = "task-1244909" [ 799.502524] env[63418]: _type = "Task" [ 799.502524] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.521156] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244909, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.522429] env[63418]: DEBUG nova.network.neutron [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updated VIF entry in instance network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.523082] env[63418]: DEBUG nova.network.neutron [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.529377] env[63418]: INFO nova.compute.manager [-] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Took 1.29 seconds to deallocate network for instance. [ 799.613279] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52083009-0e3f-bd5c-678a-2b4e39c1725b, 'name': SearchDatastore_Task, 'duration_secs': 0.013294} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.616757] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d37ca671-807e-4734-a058-9c5fc20dbde2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.622748] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 799.622748] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527ab17a-1697-48f8-2445-30a13be27bf7" [ 799.622748] env[63418]: _type = "Task" [ 799.622748] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.633410] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527ab17a-1697-48f8-2445-30a13be27bf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.830348] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26452ade-28aa-4490-86ca-ef77c25f5410 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.839799] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f818e100-40b5-47d5-8a32-06221b693708 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.872732] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8731d486-baf3-410f-8038-0c69e2995109 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.884145] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ff1b87-d7d6-4fe6-91aa-87d6c5b6dba2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.898446] env[63418]: DEBUG nova.compute.provider_tree [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.905303] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244908, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.951747] env[63418]: DEBUG oslo_concurrency.lockutils [None req-836ada57-2d4f-42c1-9909-8c6112925dd4 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "27290558-6c58-414d-bd53-c4a686890721" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.963s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.003102] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 800.004216] env[63418]: DEBUG nova.network.neutron [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.006239] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ee3e7ce1-c50d-4f0d-9d4b-ba54d48585a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.018328] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244909, 'name': Rename_Task, 'duration_secs': 0.208845} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.020119] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.020119] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 800.020119] env[63418]: value = "task-1244910" [ 800.020119] env[63418]: _type = "Task" [ 800.020119] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.020340] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2c0608b-3a19-43a9-9afa-6c10a6aba2ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.033203] env[63418]: DEBUG oslo_concurrency.lockutils [req-04abb063-c463-43c3-8f80-69efca343897 req-ea1524b4-e06c-4e8a-8b9f-9e78e47b12ec service nova] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.033796] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244910, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.037893] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.038694] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 800.038694] env[63418]: value = "task-1244911" [ 800.038694] env[63418]: _type = "Task" [ 800.038694] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.047158] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.115218] env[63418]: DEBUG nova.compute.manager [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Received event network-vif-deleted-e834b79c-1139-41ba-92e2-9875b4e19a5b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 800.115218] env[63418]: DEBUG nova.compute.manager [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Received event network-vif-plugged-48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 800.115218] env[63418]: DEBUG oslo_concurrency.lockutils [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] Acquiring lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.115556] env[63418]: DEBUG oslo_concurrency.lockutils [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.116248] env[63418]: DEBUG oslo_concurrency.lockutils [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.116248] env[63418]: DEBUG nova.compute.manager [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] No waiting events found dispatching network-vif-plugged-48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 800.117131] env[63418]: WARNING nova.compute.manager [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Received unexpected event network-vif-plugged-48c02e8f-dada-4b14-bee4-8bee405a7471 for instance with vm_state building and task_state spawning. [ 800.117131] env[63418]: DEBUG nova.compute.manager [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Received event network-changed-48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 800.117131] env[63418]: DEBUG nova.compute.manager [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Refreshing instance network info cache due to event network-changed-48c02e8f-dada-4b14-bee4-8bee405a7471. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 800.117131] env[63418]: DEBUG oslo_concurrency.lockutils [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] Acquiring lock "refresh_cache-be1c134c-f36e-4cc7-b4ef-8f30793fb4df" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.135427] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527ab17a-1697-48f8-2445-30a13be27bf7, 'name': SearchDatastore_Task, 'duration_secs': 0.010467} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.139217] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.140759] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] eaa13276-1fb1-47e2-ad1f-445bc9f4c98f/eaa13276-1fb1-47e2-ad1f-445bc9f4c98f.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.140759] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d64bb06d-12c9-4835-8e35-1ef26028d28c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.149332] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 800.149332] env[63418]: value = "task-1244912" [ 800.149332] env[63418]: _type = "Task" [ 800.149332] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.159653] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.186644] env[63418]: DEBUG nova.network.neutron [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Updating instance_info_cache with network_info: [{"id": "48c02e8f-dada-4b14-bee4-8bee405a7471", "address": "fa:16:3e:7f:d7:21", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48c02e8f-da", "ovs_interfaceid": "48c02e8f-dada-4b14-bee4-8bee405a7471", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.406151] env[63418]: DEBUG nova.scheduler.client.report [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 800.410475] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244908, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.534192] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244910, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.550066] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244911, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.664306] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244912, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.690971] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "refresh_cache-be1c134c-f36e-4cc7-b4ef-8f30793fb4df" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.691625] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Instance network_info: |[{"id": "48c02e8f-dada-4b14-bee4-8bee405a7471", "address": "fa:16:3e:7f:d7:21", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48c02e8f-da", "ovs_interfaceid": "48c02e8f-dada-4b14-bee4-8bee405a7471", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 800.692222] env[63418]: DEBUG oslo_concurrency.lockutils [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] Acquired lock "refresh_cache-be1c134c-f36e-4cc7-b4ef-8f30793fb4df" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.692497] env[63418]: DEBUG nova.network.neutron [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Refreshing network info cache for port 48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.694320] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:d7:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48c02e8f-dada-4b14-bee4-8bee405a7471', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.708016] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Creating folder: Project (4759c1d6199344c49063d35e422d6324). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.711484] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f883abb-fb6a-4b96-9568-4ddf666dabb3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.726477] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Created folder: Project (4759c1d6199344c49063d35e422d6324) in parent group-v268354. [ 800.726682] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Creating folder: Instances. Parent ref: group-v268434. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.726940] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-312581dd-a9ed-46ae-b5a1-90d0fa011946 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.740191] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Created folder: Instances in parent group-v268434. [ 800.740485] env[63418]: DEBUG oslo.service.loopingcall [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.740673] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.740891] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc2148e4-d1b7-43f1-b024-74808b590ceb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.765362] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.765362] env[63418]: value = "task-1244915" [ 800.765362] env[63418]: _type = "Task" [ 800.765362] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.775450] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244915, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.901339] env[63418]: DEBUG oslo_vmware.api [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1244908, 'name': PowerOnVM_Task, 'duration_secs': 1.041705} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.901752] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 800.902527] env[63418]: INFO nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Took 9.37 seconds to spawn the instance on the hypervisor. [ 800.902527] env[63418]: DEBUG nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.903337] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0679acf0-9dd9-43b7-8ce8-d5d6f8677867 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.916158] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.916695] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 800.919495] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.426s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.920883] env[63418]: INFO nova.compute.claims [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.927183] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "27290558-6c58-414d-bd53-c4a686890721" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.927413] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "27290558-6c58-414d-bd53-c4a686890721" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.927617] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "27290558-6c58-414d-bd53-c4a686890721-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.927794] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "27290558-6c58-414d-bd53-c4a686890721-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.927957] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "27290558-6c58-414d-bd53-c4a686890721-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.929980] env[63418]: INFO nova.compute.manager [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Terminating instance [ 800.963752] env[63418]: DEBUG nova.network.neutron [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Updated VIF entry in instance network info cache for port 48c02e8f-dada-4b14-bee4-8bee405a7471. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 800.964183] env[63418]: DEBUG nova.network.neutron [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Updating instance_info_cache with network_info: [{"id": "48c02e8f-dada-4b14-bee4-8bee405a7471", "address": "fa:16:3e:7f:d7:21", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48c02e8f-da", "ovs_interfaceid": "48c02e8f-dada-4b14-bee4-8bee405a7471", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.036291] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244910, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.050142] env[63418]: DEBUG oslo_vmware.api [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244911, 'name': PowerOnVM_Task, 'duration_secs': 0.637991} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.050416] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.050618] env[63418]: DEBUG nova.compute.manager [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.051452] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f049b72d-f433-45c0-ab62-a49a5cd3da70 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.164199] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.694133} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.164199] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] eaa13276-1fb1-47e2-ad1f-445bc9f4c98f/eaa13276-1fb1-47e2-ad1f-445bc9f4c98f.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.164199] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.164199] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cab9f051-92cc-4757-a2ee-780f8287e5b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.170182] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 801.170182] env[63418]: value = "task-1244916" [ 801.170182] env[63418]: _type = "Task" [ 801.170182] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.180782] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.199081] env[63418]: DEBUG nova.compute.manager [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 801.199081] env[63418]: DEBUG nova.compute.manager [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing instance network info cache due to event network-changed-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 801.199249] env[63418]: DEBUG oslo_concurrency.lockutils [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] Acquiring lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.199391] env[63418]: DEBUG oslo_concurrency.lockutils [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] Acquired lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.199627] env[63418]: DEBUG nova.network.neutron [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Refreshing network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.276988] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244915, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.352696] env[63418]: DEBUG nova.compute.manager [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.353516] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f18e85-c2f4-4da4-a9da-3556b2b91371 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.427772] env[63418]: DEBUG nova.compute.utils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.434588] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 801.436023] env[63418]: DEBUG nova.network.neutron [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.437512] env[63418]: DEBUG nova.compute.manager [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 801.437512] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 801.438856] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dced37-819f-4cb3-9895-15b1374de28f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.443336] env[63418]: INFO nova.compute.manager [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Took 35.15 seconds to build instance. [ 801.452877] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 801.454238] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-015ec6a5-ccc3-4da9-b471-532988e2cec6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.466079] env[63418]: DEBUG oslo_vmware.api [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 801.466079] env[63418]: value = "task-1244917" [ 801.466079] env[63418]: _type = "Task" [ 801.466079] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.471672] env[63418]: DEBUG oslo_concurrency.lockutils [req-16309bd5-f207-4d84-8f35-c4711f3bf1e0 req-f572a9e5-8481-4901-98c2-0f38b3c665fe service nova] Releasing lock "refresh_cache-be1c134c-f36e-4cc7-b4ef-8f30793fb4df" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.480176] env[63418]: DEBUG oslo_vmware.api [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.502524] env[63418]: DEBUG nova.policy [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e9d38547e714839a808ae0c7e7e97d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '381605706a8b4a1e9cca8603c800f1ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.527885] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.528167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.528745] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.528745] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.528745] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.538210] env[63418]: INFO nova.compute.manager [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Terminating instance [ 801.546130] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244910, 'name': CloneVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.573599] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.695394] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101619} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.695799] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.696755] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7eb3cd8-1b94-40ef-8f9e-f8fa208e8851 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.723073] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] eaa13276-1fb1-47e2-ad1f-445bc9f4c98f/eaa13276-1fb1-47e2-ad1f-445bc9f4c98f.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.724233] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e16e7ec-41ad-4d23-8a54-6430379cc4d3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.750871] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 801.750871] env[63418]: value = "task-1244918" [ 801.750871] env[63418]: _type = "Task" [ 801.750871] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.761712] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.776845] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244915, 'name': CreateVM_Task, 'duration_secs': 0.53118} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.777101] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.778299] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.778589] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.779222] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.779447] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df31db18-1b88-4b67-bee3-07b95b938c6d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.786167] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 801.786167] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52076ca1-6781-f67b-acbb-238eae179ee7" [ 801.786167] env[63418]: _type = "Task" [ 801.786167] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.798174] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52076ca1-6781-f67b-acbb-238eae179ee7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.866087] env[63418]: INFO nova.compute.manager [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] instance snapshotting [ 801.866364] env[63418]: WARNING nova.compute.manager [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 801.869370] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b136dd8-85d3-4cc0-bfb2-5373ca4c1153 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.889973] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ec3c58-741c-4fb5-a041-d173fe21bb6e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.935209] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 801.948584] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06b88124-bd25-44e6-b700-babc39eaa010 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.248s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.981804] env[63418]: DEBUG oslo_vmware.api [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244917, 'name': PowerOffVM_Task, 'duration_secs': 0.394271} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.983547] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.983547] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.983547] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6da92f7f-583a-4834-b26f-808b829e26af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.044050] env[63418]: DEBUG nova.compute.manager [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 802.044050] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 802.044336] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244910, 'name': CloneVM_Task, 'duration_secs': 1.594852} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.045142] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfca7935-45b9-49c1-9b07-bb1a64a6774e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.048368] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Created linked-clone VM from snapshot [ 802.049122] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b41835-2b67-4146-bc5f-3e2635d40925 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.055769] env[63418]: DEBUG nova.network.neutron [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Successfully created port: d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.060503] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 802.064290] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a709f292-96c0-427b-93e3-5507080a85e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.066744] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Uploading image 14d64c0d-64ca-4b9e-9590-07ca59045ade {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 802.078459] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 802.078459] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 802.078661] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Deleting the datastore file [datastore2] 27290558-6c58-414d-bd53-c4a686890721 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 802.080119] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-906b346e-8736-4146-9902-87dacc630d83 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.084101] env[63418]: DEBUG oslo_vmware.api [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 802.084101] env[63418]: value = "task-1244920" [ 802.084101] env[63418]: _type = "Task" [ 802.084101] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.095702] env[63418]: DEBUG oslo_vmware.api [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for the task: (returnval){ [ 802.095702] env[63418]: value = "task-1244921" [ 802.095702] env[63418]: _type = "Task" [ 802.095702] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.103033] env[63418]: DEBUG oslo_vmware.api [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.107262] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 802.107262] env[63418]: value = "vm-268433" [ 802.107262] env[63418]: _type = "VirtualMachine" [ 802.107262] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 802.107663] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2dbc6c5f-4515-4cfc-881c-d99193718e62 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.114618] env[63418]: DEBUG oslo_vmware.api [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.124410] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease: (returnval){ [ 802.124410] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fdaf7-5977-a4a0-9b12-373c68ddd771" [ 802.124410] env[63418]: _type = "HttpNfcLease" [ 802.124410] env[63418]: } obtained for exporting VM: (result){ [ 802.124410] env[63418]: value = "vm-268433" [ 802.124410] env[63418]: _type = "VirtualMachine" [ 802.124410] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 802.124410] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the lease: (returnval){ [ 802.124410] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fdaf7-5977-a4a0-9b12-373c68ddd771" [ 802.124410] env[63418]: _type = "HttpNfcLease" [ 802.124410] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 802.135510] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 802.135510] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fdaf7-5977-a4a0-9b12-373c68ddd771" [ 802.135510] env[63418]: _type = "HttpNfcLease" [ 802.135510] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 802.264616] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.294291] env[63418]: DEBUG nova.network.neutron [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updated VIF entry in instance network info cache for port a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 802.294650] env[63418]: DEBUG nova.network.neutron [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [{"id": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "address": "fa:16:3e:c5:48:53", "network": {"id": "7e925db9-82d5-49eb-8b03-061b73465d80", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-615152366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "126d2db603334c8788951d3d3acb905e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16b5ec1-ce", "ovs_interfaceid": "a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.304159] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52076ca1-6781-f67b-acbb-238eae179ee7, 'name': SearchDatastore_Task, 'duration_secs': 0.012363} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.304460] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.304685] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.304947] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.305215] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.305478] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.306412] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa798d00-d835-449d-9098-c8b9f5c6f041 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.316163] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.316373] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.317162] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba644f2-fc9c-478b-a1c4-deccee042cb8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.325030] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 802.325030] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5254a735-f5d9-9091-bb31-6f13fe386c61" [ 802.325030] env[63418]: _type = "Task" [ 802.325030] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.334376] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5254a735-f5d9-9091-bb31-6f13fe386c61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.338633] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08a91e8-ebbd-436c-9bff-b65d256f37ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.346366] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b6e2b8-4a47-4a2c-bb95-bd1c29a1de37 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.379167] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a8b328-cf94-42a8-bfb4-18b31f7758b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.387957] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ce4ee6-c0f5-476d-bac7-77abf44e3a55 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.406480] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 802.407057] env[63418]: DEBUG nova.compute.provider_tree [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 802.408846] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c5b24706-5b16-467d-b828-7d4e506c9612 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.418908] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 802.418908] env[63418]: value = "task-1244923" [ 802.418908] env[63418]: _type = "Task" [ 802.418908] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.433076] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244923, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.597378] env[63418]: DEBUG oslo_vmware.api [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244920, 'name': PowerOffVM_Task, 'duration_secs': 0.51044} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.604026] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 802.604026] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 802.604026] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3d1213c-dd6a-4d3b-a2f3-9b25a48e56b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.609688] env[63418]: DEBUG oslo_vmware.api [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Task: {'id': task-1244921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168376} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.610212] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.610563] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 802.610876] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.611202] env[63418]: INFO nova.compute.manager [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [instance: 27290558-6c58-414d-bd53-c4a686890721] Took 1.17 seconds to destroy the instance on the hypervisor. [ 802.611599] env[63418]: DEBUG oslo.service.loopingcall [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.611952] env[63418]: DEBUG nova.compute.manager [-] [instance: 27290558-6c58-414d-bd53-c4a686890721] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 802.612206] env[63418]: DEBUG nova.network.neutron [-] [instance: 27290558-6c58-414d-bd53-c4a686890721] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.634022] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 802.634022] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fdaf7-5977-a4a0-9b12-373c68ddd771" [ 802.634022] env[63418]: _type = "HttpNfcLease" [ 802.634022] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 802.634022] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 802.634022] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526fdaf7-5977-a4a0-9b12-373c68ddd771" [ 802.634022] env[63418]: _type = "HttpNfcLease" [ 802.634022] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 802.634022] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4d96bf-79e7-4910-b7f4-2d2c8241bf97 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.643923] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52449f33-edf8-6e60-c7f8-e4f052aa09db/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 802.643923] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52449f33-edf8-6e60-c7f8-e4f052aa09db/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 802.716581] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 802.717007] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 802.717220] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Deleting the datastore file [datastore1] e3dd6d5e-2b80-4724-9826-6ebfb8181c3b {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 802.718843] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-161306b3-e23c-4d63-9edc-6c5197317d11 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.726619] env[63418]: DEBUG oslo_vmware.api [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for the task: (returnval){ [ 802.726619] env[63418]: value = "task-1244925" [ 802.726619] env[63418]: _type = "Task" [ 802.726619] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.735406] env[63418]: DEBUG oslo_vmware.api [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.763123] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244918, 'name': ReconfigVM_Task, 'duration_secs': 0.84936} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.763471] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfigured VM instance instance-0000003e to attach disk [datastore2] eaa13276-1fb1-47e2-ad1f-445bc9f4c98f/eaa13276-1fb1-47e2-ad1f-445bc9f4c98f.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.764202] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-661386a5-5281-40cc-b2fa-8c2b2cd5b491 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.776681] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2218f927-c1eb-4163-9af4-a14cb8ad0306 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.778785] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 802.778785] env[63418]: value = "task-1244926" [ 802.778785] env[63418]: _type = "Task" [ 802.778785] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.792118] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244926, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.798063] env[63418]: DEBUG oslo_concurrency.lockutils [req-bdb0b179-1fb9-4bf8-a4e8-28616f5e3fa3 req-b09dc859-b016-4560-a12e-4c558e67a683 service nova] Releasing lock "refresh_cache-e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.837615] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5254a735-f5d9-9091-bb31-6f13fe386c61, 'name': SearchDatastore_Task, 'duration_secs': 0.009731} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.838485] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07269175-1357-4dfd-99f5-c8a2c7bc228f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.845123] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 802.845123] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ca3a90-6755-65a7-2232-818b8d863ec2" [ 802.845123] env[63418]: _type = "Task" [ 802.845123] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.854435] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ca3a90-6755-65a7-2232-818b8d863ec2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.936098] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244923, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.941985] env[63418]: ERROR nova.scheduler.client.report [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [req-8e31eb4d-fc1f-493f-83ff-8f3edac4b08e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8e31eb4d-fc1f-493f-83ff-8f3edac4b08e"}]} [ 802.950675] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 802.962903] env[63418]: DEBUG nova.scheduler.client.report [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 802.986510] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.986909] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.986947] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.987493] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.987746] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.988023] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.988353] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.988600] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.988838] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.989094] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.989350] env[63418]: DEBUG nova.virt.hardware [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.991039] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ea8a36-0aa3-45e6-a6a2-e6ef5d3d7120 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.994675] env[63418]: DEBUG nova.scheduler.client.report [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 802.995140] env[63418]: DEBUG nova.compute.provider_tree [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 803.005553] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6881f467-7f7b-4b7c-b4d3-128f1fe415dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.011921] env[63418]: DEBUG nova.scheduler.client.report [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 803.042796] env[63418]: DEBUG nova.scheduler.client.report [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 803.044505] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "61af37a5-8c18-47a7-817a-eb3b332b3725" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.044942] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.239402] env[63418]: DEBUG oslo_vmware.api [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Task: {'id': task-1244925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322893} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.239676] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 803.239872] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 803.240065] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.240250] env[63418]: INFO nova.compute.manager [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Took 1.20 seconds to destroy the instance on the hypervisor. [ 803.240700] env[63418]: DEBUG oslo.service.loopingcall [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.240958] env[63418]: DEBUG nova.compute.manager [-] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 803.241087] env[63418]: DEBUG nova.network.neutron [-] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 803.296796] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244926, 'name': Rename_Task, 'duration_secs': 0.194538} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.298097] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.298097] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3765f705-5c6d-44a7-a7b0-a69632d43f77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.312540] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 803.312540] env[63418]: value = "task-1244927" [ 803.312540] env[63418]: _type = "Task" [ 803.312540] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.324356] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.359826] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ca3a90-6755-65a7-2232-818b8d863ec2, 'name': SearchDatastore_Task, 'duration_secs': 0.023513} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.363360] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.363931] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] be1c134c-f36e-4cc7-b4ef-8f30793fb4df/be1c134c-f36e-4cc7-b4ef-8f30793fb4df.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.367380] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aea5e5bb-42b5-4b4d-8bfc-d565fe454083 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.377491] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 803.377491] env[63418]: value = "task-1244928" [ 803.377491] env[63418]: _type = "Task" [ 803.377491] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.389612] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244928, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.434529] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244923, 'name': CreateSnapshot_Task, 'duration_secs': 0.709133} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.435390] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 803.437239] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae92daa7-d90c-4026-beaf-f19a6c709876 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.443034] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0aaafcc-32c2-4ab0-bc82-28bba4427041 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.458954] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac98427-b68d-4ba5-9e7a-4bc0d962fb0f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.503386] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e588f67-e96b-4a06-981d-694cb82a50d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.514185] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022bfc73-69f3-476a-b265-6be568299864 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.525470] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.525890] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.536941] env[63418]: DEBUG nova.compute.provider_tree [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.548172] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 803.827783] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244927, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.828233] env[63418]: DEBUG nova.network.neutron [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Successfully updated port: d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.831057] env[63418]: DEBUG nova.network.neutron [-] [instance: 27290558-6c58-414d-bd53-c4a686890721] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.896069] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244928, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.964291] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 803.965109] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fa58420e-e43b-4eb1-9766-c8e68a245745 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.982111] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 803.982111] env[63418]: value = "task-1244929" [ 803.982111] env[63418]: _type = "Task" [ 803.982111] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.996661] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244929, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.041306] env[63418]: DEBUG nova.compute.utils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.045712] env[63418]: DEBUG nova.scheduler.client.report [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 804.078297] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.129645] env[63418]: DEBUG nova.compute.manager [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Received event network-changed-9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 804.129952] env[63418]: DEBUG nova.compute.manager [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Refreshing instance network info cache due to event network-changed-9b941608-e1b2-4bce-8249-e53c075ac132. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 804.130286] env[63418]: DEBUG oslo_concurrency.lockutils [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] Acquiring lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.130579] env[63418]: DEBUG oslo_concurrency.lockutils [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] Acquired lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.130861] env[63418]: DEBUG nova.network.neutron [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Refreshing network info cache for port 9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.242358] env[63418]: DEBUG nova.network.neutron [-] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.325275] env[63418]: DEBUG oslo_vmware.api [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1244927, 'name': PowerOnVM_Task, 'duration_secs': 0.73904} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.326107] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.326475] env[63418]: INFO nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Took 10.32 seconds to spawn the instance on the hypervisor. [ 804.327834] env[63418]: DEBUG nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 804.328844] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276fe957-e8e0-40a4-ae29-06b552f858dc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.332023] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.332023] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.332237] env[63418]: DEBUG nova.network.neutron [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.335109] env[63418]: INFO nova.compute.manager [-] [instance: 27290558-6c58-414d-bd53-c4a686890721] Took 1.72 seconds to deallocate network for instance. [ 804.356237] env[63418]: DEBUG nova.compute.manager [req-da4cd797-5171-4c05-9d3b-d7b63648f08c req-684fbf67-ec74-4fb0-846a-a0d97d047696 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Received event network-vif-plugged-d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 804.356643] env[63418]: DEBUG oslo_concurrency.lockutils [req-da4cd797-5171-4c05-9d3b-d7b63648f08c req-684fbf67-ec74-4fb0-846a-a0d97d047696 service nova] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.357189] env[63418]: DEBUG oslo_concurrency.lockutils [req-da4cd797-5171-4c05-9d3b-d7b63648f08c req-684fbf67-ec74-4fb0-846a-a0d97d047696 service nova] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.357189] env[63418]: DEBUG oslo_concurrency.lockutils [req-da4cd797-5171-4c05-9d3b-d7b63648f08c req-684fbf67-ec74-4fb0-846a-a0d97d047696 service nova] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.357300] env[63418]: DEBUG nova.compute.manager [req-da4cd797-5171-4c05-9d3b-d7b63648f08c req-684fbf67-ec74-4fb0-846a-a0d97d047696 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] No waiting events found dispatching network-vif-plugged-d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 804.357944] env[63418]: WARNING nova.compute.manager [req-da4cd797-5171-4c05-9d3b-d7b63648f08c req-684fbf67-ec74-4fb0-846a-a0d97d047696 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Received unexpected event network-vif-plugged-d0fa119b-63f3-4b54-8592-195a043ff0ee for instance with vm_state building and task_state spawning. [ 804.390545] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244928, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.720368} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.391546] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] be1c134c-f36e-4cc7-b4ef-8f30793fb4df/be1c134c-f36e-4cc7-b4ef-8f30793fb4df.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.391972] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.392597] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87c0e58c-fc03-4ee2-bea5-bde7603dcb30 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.401957] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 804.401957] env[63418]: value = "task-1244930" [ 804.401957] env[63418]: _type = "Task" [ 804.401957] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.411540] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244930, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.493914] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244929, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.552516] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.027s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.553646] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.634s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.554276] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 804.557515] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.793s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.557744] env[63418]: DEBUG nova.objects.instance [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lazy-loading 'resources' on Instance uuid 7615244b-fa19-4dbb-a851-604bb6fc2725 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.745456] env[63418]: INFO nova.compute.manager [-] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Took 1.50 seconds to deallocate network for instance. [ 804.852632] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.856161] env[63418]: INFO nova.compute.manager [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Took 35.49 seconds to build instance. [ 804.884717] env[63418]: DEBUG nova.network.neutron [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.921446] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244930, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076338} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.921853] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.922804] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3345f7a-5ff3-4b50-b099-8dfc0fe20b86 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.948834] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] be1c134c-f36e-4cc7-b4ef-8f30793fb4df/be1c134c-f36e-4cc7-b4ef-8f30793fb4df.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.949977] env[63418]: DEBUG nova.network.neutron [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Updated VIF entry in instance network info cache for port 9b941608-e1b2-4bce-8249-e53c075ac132. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.950589] env[63418]: DEBUG nova.network.neutron [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Updating instance_info_cache with network_info: [{"id": "9b941608-e1b2-4bce-8249-e53c075ac132", "address": "fa:16:3e:78:59:7e", "network": {"id": "3cda032d-334b-4784-88a8-5f0dcc78e7ea", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1114152170-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "334e6dbe6b0e4ba1917396fb73cf58c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b941608-e1", "ovs_interfaceid": "9b941608-e1b2-4bce-8249-e53c075ac132", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.955411] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f5bb6f1-b695-44cd-9608-c78b600c8de6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.977951] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 804.977951] env[63418]: value = "task-1244931" [ 804.977951] env[63418]: _type = "Task" [ 804.977951] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.991835] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244931, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.997946] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244929, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.061415] env[63418]: DEBUG nova.compute.utils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.067678] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 805.067678] env[63418]: DEBUG nova.network.neutron [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.107408] env[63418]: DEBUG nova.network.neutron [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.119156] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.119624] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.119624] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.119831] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.120613] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.123390] env[63418]: DEBUG nova.policy [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc5c0fadf0b54b71957a87d56ea10acc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c828658cd364400a9207d0c90f3ae1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.125971] env[63418]: INFO nova.compute.manager [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Terminating instance [ 805.253161] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.358842] env[63418]: DEBUG oslo_concurrency.lockutils [None req-955c5e7b-4e14-4d9f-8232-ea56874cc713 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.570s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.423704] env[63418]: DEBUG nova.network.neutron [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Successfully created port: 2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.447985] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd31d8c-4437-40ec-abfc-f58de068e3a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.454375] env[63418]: DEBUG oslo_concurrency.lockutils [req-b6bb649b-8e85-48db-8aaf-e686a55cd4fa req-6683f5e8-087e-4abe-8e08-c56f3d70c582 service nova] Releasing lock "refresh_cache-7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.458861] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d907a363-3899-4677-9ffb-1198c6da51b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.504584] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512e6cb0-d93a-4ee9-a528-e69bb84b1a87 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.513518] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244931, 'name': ReconfigVM_Task, 'duration_secs': 0.517223} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.514344] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Reconfigured VM instance instance-0000003f to attach disk [datastore1] be1c134c-f36e-4cc7-b4ef-8f30793fb4df/be1c134c-f36e-4cc7-b4ef-8f30793fb4df.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.515047] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9cb946a8-2017-4be2-a63e-0be399781a72 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.525752] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 805.525752] env[63418]: value = "task-1244932" [ 805.525752] env[63418]: _type = "Task" [ 805.525752] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.530939] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244929, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.535722] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19131af8-d2b6-4e3a-9c48-3d42b0b4a253 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.549667] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244932, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.558667] env[63418]: DEBUG nova.compute.provider_tree [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.568176] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 805.613328] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.614017] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Instance network_info: |[{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 805.614447] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:0e:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd0fa119b-63f3-4b54-8592-195a043ff0ee', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.624351] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating folder: Project (381605706a8b4a1e9cca8603c800f1ac). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.624351] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa59206f-b598-48dd-9874-cd518b2e48ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.637463] env[63418]: DEBUG nova.compute.manager [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 805.637731] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 805.638304] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.638526] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.638737] env[63418]: INFO nova.compute.manager [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Attaching volume b91a8528-5ded-45b1-b1bb-31b41d7fff4b to /dev/sdb [ 805.642155] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50f0d13-97b0-4f61-8b5a-5bb7d2f219a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.645595] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created folder: Project (381605706a8b4a1e9cca8603c800f1ac) in parent group-v268354. [ 805.645595] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating folder: Instances. Parent ref: group-v268439. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.647991] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06333343-9631-42a0-b84b-76134459c650 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.655188] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.655466] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bda9fab6-c764-4d5f-8723-b585796c2192 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.664023] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created folder: Instances in parent group-v268439. [ 805.664023] env[63418]: DEBUG oslo.service.loopingcall [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.664023] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.664023] env[63418]: DEBUG oslo_vmware.api [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 805.664023] env[63418]: value = "task-1244935" [ 805.664023] env[63418]: _type = "Task" [ 805.664023] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.664023] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bec1548-788b-48d4-b961-1f66c74a354e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.687674] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d343e05c-a5ee-46ee-aa8b-c5f0eaec53b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.697132] env[63418]: DEBUG oslo_vmware.api [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.699840] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.699840] env[63418]: value = "task-1244936" [ 805.699840] env[63418]: _type = "Task" [ 805.699840] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.700683] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c068048-1c5c-4ec2-862d-e7698b2174ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.716446] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244936, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.721125] env[63418]: DEBUG nova.virt.block_device [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updating existing volume attachment record: 112681d4-c188-46d4-a1e5-872f4f3527b5 {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 806.010815] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244929, 'name': CloneVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.047330] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244932, 'name': Rename_Task, 'duration_secs': 0.241579} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.047932] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.048379] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24233711-1b1e-416e-9d36-85fcd7f07f6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.061022] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 806.061022] env[63418]: value = "task-1244939" [ 806.061022] env[63418]: _type = "Task" [ 806.061022] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.064849] env[63418]: DEBUG nova.scheduler.client.report [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 806.083914] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.195740] env[63418]: DEBUG oslo_vmware.api [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244935, 'name': PowerOffVM_Task, 'duration_secs': 0.385262} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.196153] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 806.196390] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 806.196682] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-172d41e2-bee0-437c-9767-ebce1c867ca8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.216389] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244936, 'name': CreateVM_Task, 'duration_secs': 0.509684} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.216568] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.217352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.217530] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.217870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.219330] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4375089-3f1e-4e88-a137-ecd9117ef952 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.224095] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.224393] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.230281] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 806.230281] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cffdab-839f-2ae9-efc8-aa173dfb71cd" [ 806.230281] env[63418]: _type = "Task" [ 806.230281] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.243893] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cffdab-839f-2ae9-efc8-aa173dfb71cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.282154] env[63418]: DEBUG nova.compute.manager [req-2a583795-a92e-4da6-bff1-64b5951ba91a req-1865dc49-d984-4cb1-8b20-c2b4026c6220 service nova] [instance: 27290558-6c58-414d-bd53-c4a686890721] Received event network-vif-deleted-f161fb5a-0ae6-4123-8922-c602842c4933 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 806.386920] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 806.387240] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 806.388151] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore2] 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 806.388522] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9c1da7c-574c-4694-a7c7-fcc40f6d4fbb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.396356] env[63418]: DEBUG nova.compute.manager [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Received event network-vif-deleted-a16b5ec1-ceb0-49e9-bfd7-66f618d2a6a6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 806.396598] env[63418]: DEBUG nova.compute.manager [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Received event network-changed-d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 806.396666] env[63418]: DEBUG nova.compute.manager [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Refreshing instance network info cache due to event network-changed-d0fa119b-63f3-4b54-8592-195a043ff0ee. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 806.396894] env[63418]: DEBUG oslo_concurrency.lockutils [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.397052] env[63418]: DEBUG oslo_concurrency.lockutils [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.397232] env[63418]: DEBUG nova.network.neutron [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Refreshing network info cache for port d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.401322] env[63418]: DEBUG oslo_vmware.api [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 806.401322] env[63418]: value = "task-1244942" [ 806.401322] env[63418]: _type = "Task" [ 806.401322] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.427180] env[63418]: DEBUG oslo_vmware.api [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.510490] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244929, 'name': CloneVM_Task, 'duration_secs': 2.500484} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.510764] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Created linked-clone VM from snapshot [ 806.511677] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6284870e-5689-4cf1-afff-c6feacd424b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.521618] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Uploading image 73126df0-d2dd-40c7-9e81-e5d195e1d214 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 806.544882] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 806.544882] env[63418]: value = "vm-268438" [ 806.544882] env[63418]: _type = "VirtualMachine" [ 806.544882] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 806.545207] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9eed87a4-290c-41d1-bcf4-dd1b03aed303 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.553997] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease: (returnval){ [ 806.553997] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc9ecf-6c17-c690-6ff0-be57e37f8b30" [ 806.553997] env[63418]: _type = "HttpNfcLease" [ 806.553997] env[63418]: } obtained for exporting VM: (result){ [ 806.553997] env[63418]: value = "vm-268438" [ 806.553997] env[63418]: _type = "VirtualMachine" [ 806.553997] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 806.554317] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the lease: (returnval){ [ 806.554317] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc9ecf-6c17-c690-6ff0-be57e37f8b30" [ 806.554317] env[63418]: _type = "HttpNfcLease" [ 806.554317] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 806.567330] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 806.567330] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc9ecf-6c17-c690-6ff0-be57e37f8b30" [ 806.567330] env[63418]: _type = "HttpNfcLease" [ 806.567330] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 806.575168] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.017s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.578100] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244939, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.578717] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.819s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.580527] env[63418]: INFO nova.compute.claims [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.586765] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 806.605212] env[63418]: INFO nova.scheduler.client.report [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleted allocations for instance 7615244b-fa19-4dbb-a851-604bb6fc2725 [ 806.626526] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.626914] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.627101] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.627323] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.627510] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.627667] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.628064] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.628555] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.628856] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.629833] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.629833] env[63418]: DEBUG nova.virt.hardware [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.631581] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6d7480-6d1b-49dc-bc09-fcc3f332ca2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.643274] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd74755-9aba-4547-b18b-db8149f12bf1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.727654] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 806.742679] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cffdab-839f-2ae9-efc8-aa173dfb71cd, 'name': SearchDatastore_Task, 'duration_secs': 0.013477} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.743187] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.743549] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.743925] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.744157] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.744364] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.744715] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fae7f84-5eb0-4c18-9167-990d4093ccff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.757022] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.757022] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.757022] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9095b5f-eae5-416c-80ab-4f0300a2e40b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.763187] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 806.763187] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523988de-8e5c-5c70-b260-cd1e612d3bcb" [ 806.763187] env[63418]: _type = "Task" [ 806.763187] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.773790] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523988de-8e5c-5c70-b260-cd1e612d3bcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.922165] env[63418]: DEBUG oslo_vmware.api [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1244942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210134} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.922165] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.922165] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 806.922165] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.922165] env[63418]: INFO nova.compute.manager [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Took 1.28 seconds to destroy the instance on the hypervisor. [ 806.922511] env[63418]: DEBUG oslo.service.loopingcall [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.922511] env[63418]: DEBUG nova.compute.manager [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 806.922511] env[63418]: DEBUG nova.network.neutron [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.067671] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 807.067671] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc9ecf-6c17-c690-6ff0-be57e37f8b30" [ 807.067671] env[63418]: _type = "HttpNfcLease" [ 807.067671] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 807.069022] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 807.069022] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cc9ecf-6c17-c690-6ff0-be57e37f8b30" [ 807.069022] env[63418]: _type = "HttpNfcLease" [ 807.069022] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 807.069022] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9904f46e-04b8-4300-ba69-14aabd514aa2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.077183] env[63418]: DEBUG oslo_vmware.api [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244939, 'name': PowerOnVM_Task, 'duration_secs': 0.674864} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.081017] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.081017] env[63418]: INFO nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Took 8.62 seconds to spawn the instance on the hypervisor. [ 807.081017] env[63418]: DEBUG nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 807.081017] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4facd6d2-7a90-44c1-9ea5-6ac237abff24 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.089275] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248454d-4c73-737a-613a-bf9421151849/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 807.089709] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248454d-4c73-737a-613a-bf9421151849/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 807.167167] env[63418]: DEBUG nova.network.neutron [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Successfully updated port: 2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.176241] env[63418]: DEBUG oslo_concurrency.lockutils [None req-658c9286-4193-4b01-81f7-df30e39a81ce tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "7615244b-fa19-4dbb-a851-604bb6fc2725" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.071s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.243027] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-19f7cd65-82a8-41c7-b392-f413cdf6ff49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.252826] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.275331] env[63418]: DEBUG nova.network.neutron [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updated VIF entry in instance network info cache for port d0fa119b-63f3-4b54-8592-195a043ff0ee. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 807.275331] env[63418]: DEBUG nova.network.neutron [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.287295] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523988de-8e5c-5c70-b260-cd1e612d3bcb, 'name': SearchDatastore_Task, 'duration_secs': 0.014524} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.288796] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce441910-589a-41ab-bac1-da0160384dd8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.297926] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 807.297926] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b497d-a5d5-ea17-aaee-0ada8c31aac9" [ 807.297926] env[63418]: _type = "Task" [ 807.297926] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.309549] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b497d-a5d5-ea17-aaee-0ada8c31aac9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.676055] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.676785] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.676785] env[63418]: DEBUG nova.network.neutron [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.694680] env[63418]: INFO nova.compute.manager [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Took 36.19 seconds to build instance. [ 807.781540] env[63418]: DEBUG oslo_concurrency.lockutils [req-2f97e229-7d76-4d12-b75f-ddf0af2a611e req-0b3d8cc7-65d9-46d2-b727-50da261ea679 service nova] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.815826] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b497d-a5d5-ea17-aaee-0ada8c31aac9, 'name': SearchDatastore_Task, 'duration_secs': 0.027526} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.816164] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.816483] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a/7b0c70aa-e2bc-4131-97b4-4e53a378940a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.819422] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c442e898-5d2d-4f6e-976e-2149ae06d8fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.829944] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 807.829944] env[63418]: value = "task-1244944" [ 807.829944] env[63418]: _type = "Task" [ 807.829944] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.834223] env[63418]: DEBUG nova.network.neutron [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.846394] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.963620] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc752f6d-8e1f-43f7-b0c6-b371b9d27d09 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.976054] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896057fd-ba21-4254-80fa-e30c8c393386 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.020223] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f48bd1f-fd75-4baa-b3a5-51dfcdc6cba9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.030289] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655d9aea-0cc8-4c93-997b-ab9cf856c9b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.049381] env[63418]: DEBUG nova.compute.provider_tree [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.199603] env[63418]: DEBUG oslo_concurrency.lockutils [None req-107b5efe-5c7b-406c-a4a4-72b8f0e2083c tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.728s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.257967] env[63418]: DEBUG nova.network.neutron [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.340208] env[63418]: INFO nova.compute.manager [-] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Took 1.42 seconds to deallocate network for instance. [ 808.347081] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244944, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.553249] env[63418]: DEBUG nova.scheduler.client.report [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 808.601869] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "0c151b0c-c383-4421-b506-b7afa95e2072" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.602319] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "0c151b0c-c383-4421-b506-b7afa95e2072" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.602604] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "0c151b0c-c383-4421-b506-b7afa95e2072-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.602845] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "0c151b0c-c383-4421-b506-b7afa95e2072-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.603073] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "0c151b0c-c383-4421-b506-b7afa95e2072-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.607892] env[63418]: INFO nova.compute.manager [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Terminating instance [ 808.614747] env[63418]: DEBUG nova.compute.manager [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-changed-1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 808.614954] env[63418]: DEBUG nova.compute.manager [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing instance network info cache due to event network-changed-1eca951d-2df6-4179-aa0c-6babf13838f3. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 808.616258] env[63418]: DEBUG oslo_concurrency.lockutils [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.616258] env[63418]: DEBUG oslo_concurrency.lockutils [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.616258] env[63418]: DEBUG nova.network.neutron [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing network info cache for port 1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.629877] env[63418]: DEBUG nova.network.neutron [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [{"id": "2024d711-87c3-497d-8373-842a53f06798", "address": "fa:16:3e:e9:da:bf", "network": {"id": "3c73f0ee-1a86-4875-8c44-5658ef2ff43f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1851341556-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c828658cd364400a9207d0c90f3ae1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2024d711-87", "ovs_interfaceid": "2024d711-87c3-497d-8373-842a53f06798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.695969] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "f632d71d-498f-4914-9895-8f37187a295f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.696348] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "f632d71d-498f-4914-9895-8f37187a295f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.696573] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "f632d71d-498f-4914-9895-8f37187a295f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.696788] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "f632d71d-498f-4914-9895-8f37187a295f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.698019] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "f632d71d-498f-4914-9895-8f37187a295f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.700496] env[63418]: INFO nova.compute.manager [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Terminating instance [ 808.703958] env[63418]: DEBUG nova.compute.manager [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Received event network-vif-plugged-2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 808.704190] env[63418]: DEBUG oslo_concurrency.lockutils [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] Acquiring lock "fa4c0315-1ef5-4491-ab18-d49563b778fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.704550] env[63418]: DEBUG oslo_concurrency.lockutils [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.704760] env[63418]: DEBUG oslo_concurrency.lockutils [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.705024] env[63418]: DEBUG nova.compute.manager [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] No waiting events found dispatching network-vif-plugged-2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 808.705177] env[63418]: WARNING nova.compute.manager [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Received unexpected event network-vif-plugged-2024d711-87c3-497d-8373-842a53f06798 for instance with vm_state building and task_state spawning. [ 808.705316] env[63418]: DEBUG nova.compute.manager [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Received event network-changed-2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 808.705473] env[63418]: DEBUG nova.compute.manager [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Refreshing instance network info cache due to event network-changed-2024d711-87c3-497d-8373-842a53f06798. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 808.705641] env[63418]: DEBUG oslo_concurrency.lockutils [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] Acquiring lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.843570] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.663426} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.843744] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a/7b0c70aa-e2bc-4131-97b4-4e53a378940a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 808.843929] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.844243] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60fdd82c-0d57-4bf9-a26b-5cd7fd57cad3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.853685] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.854179] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 808.854179] env[63418]: value = "task-1244946" [ 808.854179] env[63418]: _type = "Task" [ 808.854179] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.863436] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.062068] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.064079] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.031s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.065828] env[63418]: INFO nova.compute.claims [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.112696] env[63418]: DEBUG nova.compute.manager [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 809.113011] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.113971] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93f3ca8-a303-43b2-bb48-c38a9c9c2cbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.123374] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.123681] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f107ba15-c1ab-4221-8fb0-7a6d05f88481 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.132525] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Releasing lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.132665] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Instance network_info: |[{"id": "2024d711-87c3-497d-8373-842a53f06798", "address": "fa:16:3e:e9:da:bf", "network": {"id": "3c73f0ee-1a86-4875-8c44-5658ef2ff43f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1851341556-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c828658cd364400a9207d0c90f3ae1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2024d711-87", "ovs_interfaceid": "2024d711-87c3-497d-8373-842a53f06798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 809.134393] env[63418]: DEBUG oslo_concurrency.lockutils [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] Acquired lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.134657] env[63418]: DEBUG nova.network.neutron [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Refreshing network info cache for port 2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.136088] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:da:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ded8bac-871f-491b-94ec-cb67c08bc828', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2024d711-87c3-497d-8373-842a53f06798', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.147030] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Creating folder: Project (1c828658cd364400a9207d0c90f3ae1c). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.147030] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 809.147030] env[63418]: value = "task-1244947" [ 809.147030] env[63418]: _type = "Task" [ 809.147030] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.147030] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-531058c7-75c1-43d7-b9f8-2b3d9a9a5095 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.158747] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244947, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.165582] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Created folder: Project (1c828658cd364400a9207d0c90f3ae1c) in parent group-v268354. [ 809.165854] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Creating folder: Instances. Parent ref: group-v268444. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.166101] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b856c2be-21ce-4f75-b427-c35445abb5fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.182480] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Created folder: Instances in parent group-v268444. [ 809.182745] env[63418]: DEBUG oslo.service.loopingcall [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.182961] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.183239] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19b719d0-797f-4029-a2ca-a1b3c87638e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.206862] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.206862] env[63418]: value = "task-1244950" [ 809.206862] env[63418]: _type = "Task" [ 809.206862] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.211562] env[63418]: DEBUG nova.compute.manager [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 809.211884] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.212935] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d431a5-3b16-4fa4-b5fe-de43ebc4c362 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.222407] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.225847] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.226261] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77d2c4a5-27cb-4e02-ab7c-6dabd32a1cf6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.237168] env[63418]: DEBUG oslo_vmware.api [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 809.237168] env[63418]: value = "task-1244951" [ 809.237168] env[63418]: _type = "Task" [ 809.237168] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.248105] env[63418]: DEBUG oslo_vmware.api [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.370335] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.540645] env[63418]: DEBUG nova.network.neutron [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updated VIF entry in instance network info cache for port 1eca951d-2df6-4179-aa0c-6babf13838f3. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.541203] env[63418]: DEBUG nova.network.neutron [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.564801] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "bd3a76a1-8066-4ac4-8aa8-449af82d52c0" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.565214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "bd3a76a1-8066-4ac4-8aa8-449af82d52c0" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.659931] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244947, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.718941] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.748956] env[63418]: DEBUG oslo_vmware.api [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244951, 'name': PowerOffVM_Task, 'duration_secs': 0.425013} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.752180] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.752392] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.752703] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00ef5f6c-6e53-44fd-b9a8-8029128270ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.867743] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.741922} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.868032] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.868849] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4607a6-7778-46ad-a035-7fbc2996aac9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.893567] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a/7b0c70aa-e2bc-4131-97b4-4e53a378940a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.893976] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-642f633f-aec1-4a1a-bcea-edaa0a89837e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.918496] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 809.918496] env[63418]: value = "task-1244953" [ 809.918496] env[63418]: _type = "Task" [ 809.918496] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.930378] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244953, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.963221] env[63418]: DEBUG nova.network.neutron [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updated VIF entry in instance network info cache for port 2024d711-87c3-497d-8373-842a53f06798. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.963661] env[63418]: DEBUG nova.network.neutron [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [{"id": "2024d711-87c3-497d-8373-842a53f06798", "address": "fa:16:3e:e9:da:bf", "network": {"id": "3c73f0ee-1a86-4875-8c44-5658ef2ff43f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1851341556-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c828658cd364400a9207d0c90f3ae1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2024d711-87", "ovs_interfaceid": "2024d711-87c3-497d-8373-842a53f06798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.044989] env[63418]: DEBUG oslo_concurrency.lockutils [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.045377] env[63418]: DEBUG nova.compute.manager [req-b73232c1-88a6-40dd-8233-8916365223a4 req-897f9c08-dd18-4f0e-b895-f504941f79f9 service nova] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Received event network-vif-deleted-f7766654-bf8e-4a36-b935-da5442b28a70 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 810.068294] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "bd3a76a1-8066-4ac4-8aa8-449af82d52c0" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.068897] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 810.158715] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244947, 'name': PowerOffVM_Task, 'duration_secs': 0.529623} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.161397] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.161581] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.162851] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-906c1b82-f434-4591-85cb-6b683da67cd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.222750] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.278048] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 810.278411] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268443', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'name': 'volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e7b9a60-9e4b-46e3-adf9-7337dc7d2171', 'attached_at': '', 'detached_at': '', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'serial': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 810.279346] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6162135-bf53-40a0-a456-c0faa11ae006 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.301499] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023e1f64-bc8d-4db4-beef-e5e1315a8e7d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.328531] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b/volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.331477] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12b5f686-0230-4e85-abe8-fee3956282f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.351954] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 810.351954] env[63418]: value = "task-1244955" [ 810.351954] env[63418]: _type = "Task" [ 810.351954] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.363500] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244955, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.433390] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.438723] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bc7898-ea17-4c66-95ea-92d2dfa65a6c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.447969] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b70d629-990a-4aa9-aa89-f728007f8e40 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.481421] env[63418]: DEBUG oslo_concurrency.lockutils [req-7bda5302-9c36-44fa-adc9-e38c386ba20e req-6667a94f-aaac-4d54-9a99-87bab3e8658b service nova] Releasing lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.482676] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816664d9-eb83-40d0-870f-e7c1c5cdf794 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.491668] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a0cbcc-acad-4a58-b14f-35705a87605c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.507129] env[63418]: DEBUG nova.compute.provider_tree [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.574344] env[63418]: DEBUG nova.compute.utils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.576080] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 810.576328] env[63418]: DEBUG nova.network.neutron [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.640942] env[63418]: DEBUG nova.policy [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63a9aca16c834b4b8a8688b2aa819235', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8b28e6711f1b4b948dbfc124eca1e528', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.720362] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.865174] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244955, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.930730] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244953, 'name': ReconfigVM_Task, 'duration_secs': 0.677359} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.931144] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a/7b0c70aa-e2bc-4131-97b4-4e53a378940a.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.931842] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36eb5163-75e2-4947-bf30-100d7d685548 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.942122] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 810.942122] env[63418]: value = "task-1244956" [ 810.942122] env[63418]: _type = "Task" [ 810.942122] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.952690] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244956, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.005613] env[63418]: DEBUG nova.network.neutron [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Successfully created port: e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.034767] env[63418]: ERROR nova.scheduler.client.report [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [req-82087e55-67a8-4989-8e18-69096edd964a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-82087e55-67a8-4989-8e18-69096edd964a"}]} [ 811.056419] env[63418]: DEBUG nova.scheduler.client.report [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 811.075069] env[63418]: DEBUG nova.scheduler.client.report [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 811.075333] env[63418]: DEBUG nova.compute.provider_tree [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.079296] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 811.088720] env[63418]: DEBUG nova.scheduler.client.report [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 811.109981] env[63418]: DEBUG nova.scheduler.client.report [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 811.220047] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.365668] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244955, 'name': ReconfigVM_Task, 'duration_secs': 0.719001} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.368764] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Reconfigured VM instance instance-00000033 to attach disk [datastore1] volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b/volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.374392] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-585ac2fe-0789-4822-9257-35ff4da8a15e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.395208] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 811.395208] env[63418]: value = "task-1244957" [ 811.395208] env[63418]: _type = "Task" [ 811.395208] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.409184] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.426801] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f4bcc7-e3ac-447b-a2a1-4dd7a02da22f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.438204] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a42ebc4-5d0e-4366-8d98-90c8302bea68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.479501] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e6baf1-bb43-4003-9335-7c940921aaf7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.490610] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244956, 'name': Rename_Task, 'duration_secs': 0.252346} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.491850] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5a095f-46b4-42ef-ae80-4e3f2dda462e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.496200] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.496336] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-109e5bba-31c0-4cca-9947-8220725e58a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.509161] env[63418]: DEBUG nova.compute.provider_tree [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.512296] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 811.512296] env[63418]: value = "task-1244958" [ 811.512296] env[63418]: _type = "Task" [ 811.512296] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.524142] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244958, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.720648] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.909201] env[63418]: DEBUG oslo_vmware.api [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244957, 'name': ReconfigVM_Task, 'duration_secs': 0.186147} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.909615] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268443', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'name': 'volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e7b9a60-9e4b-46e3-adf9-7337dc7d2171', 'attached_at': '', 'detached_at': '', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'serial': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 812.027443] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244958, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.057145] env[63418]: DEBUG nova.scheduler.client.report [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 812.057600] env[63418]: DEBUG nova.compute.provider_tree [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 77 to 78 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 812.057931] env[63418]: DEBUG nova.compute.provider_tree [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 812.088980] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 812.221532] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.527653] env[63418]: DEBUG oslo_vmware.api [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1244958, 'name': PowerOnVM_Task, 'duration_secs': 0.696143} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.527954] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.528403] env[63418]: INFO nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 9.58 seconds to spawn the instance on the hypervisor. [ 812.528653] env[63418]: DEBUG nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 812.529503] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b520ca-0368-4a6d-a356-607e498a90f9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.565800] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.502s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.566366] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 812.569044] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.101s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.570579] env[63418]: INFO nova.compute.claims [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.722910] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.881086] env[63418]: DEBUG nova.network.neutron [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Successfully updated port: e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.942738] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 812.943182] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 812.943497] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleting the datastore file [datastore1] f632d71d-498f-4914-9895-8f37187a295f {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.944050] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd63a13e-f2e0-4b75-8816-c74876265564 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.952144] env[63418]: DEBUG oslo_vmware.api [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 812.952144] env[63418]: value = "task-1244959" [ 812.952144] env[63418]: _type = "Task" [ 812.952144] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.953033] env[63418]: DEBUG nova.objects.instance [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lazy-loading 'flavor' on Instance uuid 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.966637] env[63418]: DEBUG oslo_vmware.api [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244959, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.971097] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 812.971367] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 812.971729] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleting the datastore file [datastore2] 0c151b0c-c383-4421-b506-b7afa95e2072 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.972692] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-953d09fd-c055-44e0-bff8-481442fd67af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.980266] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for the task: (returnval){ [ 812.980266] env[63418]: value = "task-1244960" [ 812.980266] env[63418]: _type = "Task" [ 812.980266] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.988943] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.049269] env[63418]: INFO nova.compute.manager [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 29.54 seconds to build instance. [ 813.075386] env[63418]: DEBUG nova.compute.utils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.079870] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 813.079870] env[63418]: DEBUG nova.network.neutron [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.129195] env[63418]: DEBUG nova.policy [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d29f8428db741c0a6c9f6f6dfdf50d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0462b212fa4449c2a6f98cec2f186f51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.227540] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244950, 'name': CreateVM_Task, 'duration_secs': 4.012732} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.227740] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.229283] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.229368] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.229815] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.230144] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65f28e5f-c893-4847-b4ec-2fcd0f9052b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.237113] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 813.237113] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520fbad7-d9b7-bc40-f153-0446ec145fe7" [ 813.237113] env[63418]: _type = "Task" [ 813.237113] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.249575] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520fbad7-d9b7-bc40-f153-0446ec145fe7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.383169] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "refresh_cache-f7554ae4-c7a7-4111-a830-10f9029dc074" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.383337] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquired lock "refresh_cache-f7554ae4-c7a7-4111-a830-10f9029dc074" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.383498] env[63418]: DEBUG nova.network.neutron [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.395170] env[63418]: DEBUG nova.network.neutron [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Successfully created port: f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.465190] env[63418]: DEBUG oslo_vmware.api [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244959, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291815} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.465694] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c4aec05f-dc5d-48d8-aac6-10b12be618f4 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.827s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.466723] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.466835] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 813.467018] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 813.467195] env[63418]: INFO nova.compute.manager [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: f632d71d-498f-4914-9895-8f37187a295f] Took 4.26 seconds to destroy the instance on the hypervisor. [ 813.467432] env[63418]: DEBUG oslo.service.loopingcall [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.467790] env[63418]: DEBUG nova.compute.manager [-] [instance: f632d71d-498f-4914-9895-8f37187a295f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 813.467889] env[63418]: DEBUG nova.network.neutron [-] [instance: f632d71d-498f-4914-9895-8f37187a295f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.491081] env[63418]: DEBUG oslo_vmware.api [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Task: {'id': task-1244960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264918} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.491486] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.491699] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 813.491904] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 813.492072] env[63418]: INFO nova.compute.manager [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Took 4.38 seconds to destroy the instance on the hypervisor. [ 813.492321] env[63418]: DEBUG oslo.service.loopingcall [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.492512] env[63418]: DEBUG nova.compute.manager [-] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 813.492602] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.551896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-80780460-2fb5-49ee-99e1-8ca49b05ebee tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.387s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.580061] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 813.597045] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.597045] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.597045] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.597045] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.597401] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.597401] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.597401] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.597401] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.597401] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.597548] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.597548] env[63418]: DEBUG nova.virt.hardware [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.602031] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd620bed-16f9-4742-bd5e-c8e9c550daec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.609467] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac05a2e-411a-4b63-8fce-a69014be26ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.751314] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520fbad7-d9b7-bc40-f153-0446ec145fe7, 'name': SearchDatastore_Task, 'duration_secs': 0.028514} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.753525] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.753770] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.754059] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.754225] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.754532] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.754857] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bfbbf51-ec15-4273-888e-8e6bc6bf79d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.769525] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.769647] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.770409] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09291276-69ea-40e1-b81f-62e5bacd058a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.776956] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 813.776956] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a4ca55-743c-d21c-25a3-214496e35004" [ 813.776956] env[63418]: _type = "Task" [ 813.776956] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.789350] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a4ca55-743c-d21c-25a3-214496e35004, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.914240] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52449f33-edf8-6e60-c7f8-e4f052aa09db/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 813.915817] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f85ba5-220a-46e8-90d1-b1d611ae3465 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.926921] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52449f33-edf8-6e60-c7f8-e4f052aa09db/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 813.927189] env[63418]: ERROR oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52449f33-edf8-6e60-c7f8-e4f052aa09db/disk-0.vmdk due to incomplete transfer. [ 813.928546] env[63418]: DEBUG nova.network.neutron [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.929753] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-18e66cf2-65db-49c1-9f70-ab02c9f5dcc3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.933854] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0a4c9b-8c9d-4cf0-9856-ef5aa2b7ca3e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.939584] env[63418]: DEBUG oslo_vmware.rw_handles [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52449f33-edf8-6e60-c7f8-e4f052aa09db/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 813.939841] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Uploaded image 14d64c0d-64ca-4b9e-9590-07ca59045ade to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 813.942222] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 813.944665] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3ae6887b-14fb-453d-8d86-8b78a14b671d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.951156] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd110eec-1a91-4647-bed6-8c855d0b0e10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.960959] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 813.960959] env[63418]: value = "task-1244961" [ 813.960959] env[63418]: _type = "Task" [ 813.960959] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.996965] env[63418]: DEBUG nova.compute.manager [req-9af606db-dc92-4728-aae9-b249e2a18612 req-e7bb5b89-43c3-4058-b91d-97f739d151b7 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Received event network-vif-plugged-e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 813.997148] env[63418]: DEBUG oslo_concurrency.lockutils [req-9af606db-dc92-4728-aae9-b249e2a18612 req-e7bb5b89-43c3-4058-b91d-97f739d151b7 service nova] Acquiring lock "f7554ae4-c7a7-4111-a830-10f9029dc074-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.997361] env[63418]: DEBUG oslo_concurrency.lockutils [req-9af606db-dc92-4728-aae9-b249e2a18612 req-e7bb5b89-43c3-4058-b91d-97f739d151b7 service nova] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.997528] env[63418]: DEBUG oslo_concurrency.lockutils [req-9af606db-dc92-4728-aae9-b249e2a18612 req-e7bb5b89-43c3-4058-b91d-97f739d151b7 service nova] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.997696] env[63418]: DEBUG nova.compute.manager [req-9af606db-dc92-4728-aae9-b249e2a18612 req-e7bb5b89-43c3-4058-b91d-97f739d151b7 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] No waiting events found dispatching network-vif-plugged-e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 813.997853] env[63418]: WARNING nova.compute.manager [req-9af606db-dc92-4728-aae9-b249e2a18612 req-e7bb5b89-43c3-4058-b91d-97f739d151b7 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Received unexpected event network-vif-plugged-e1a45abb-8278-47dc-9780-4592d326c13d for instance with vm_state building and task_state spawning. [ 813.999192] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccfb7ae-354e-4414-909f-496fdb946ed8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.006243] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244961, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.012657] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4a6009-8d62-4a3d-be44-a3fe412dda4b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.027801] env[63418]: DEBUG nova.compute.provider_tree [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.106688] env[63418]: DEBUG nova.network.neutron [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Updating instance_info_cache with network_info: [{"id": "e1a45abb-8278-47dc-9780-4592d326c13d", "address": "fa:16:3e:c5:d7:f5", "network": {"id": "67e9ab9a-8452-4b02-b84c-1fa4d4f18035", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-822446023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b28e6711f1b4b948dbfc124eca1e528", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45abb-82", "ovs_interfaceid": "e1a45abb-8278-47dc-9780-4592d326c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.287611] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a4ca55-743c-d21c-25a3-214496e35004, 'name': SearchDatastore_Task, 'duration_secs': 0.024352} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.288450] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9114199e-cfc1-40bc-9dde-93836c9e1335 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.294384] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 814.294384] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526ecb96-5e33-a263-ea16-54dde4bd1cc7" [ 814.294384] env[63418]: _type = "Task" [ 814.294384] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.304720] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526ecb96-5e33-a263-ea16-54dde4bd1cc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.398926] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.409058] env[63418]: DEBUG nova.network.neutron [-] [instance: f632d71d-498f-4914-9895-8f37187a295f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.473161] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244961, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.560723] env[63418]: DEBUG nova.scheduler.client.report [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 814.561076] env[63418]: DEBUG nova.compute.provider_tree [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 78 to 79 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 814.561202] env[63418]: DEBUG nova.compute.provider_tree [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.587953] env[63418]: DEBUG nova.compute.manager [req-aa55f70f-0e60-49f6-b974-34e556dc27a0 req-880f251a-827a-4a18-ba15-1c4e320f1ea0 service nova] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Received event network-vif-deleted-881c7ce8-24ab-4025-90bb-9d5c0bc92bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 814.595199] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 814.610551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Releasing lock "refresh_cache-f7554ae4-c7a7-4111-a830-10f9029dc074" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.610952] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Instance network_info: |[{"id": "e1a45abb-8278-47dc-9780-4592d326c13d", "address": "fa:16:3e:c5:d7:f5", "network": {"id": "67e9ab9a-8452-4b02-b84c-1fa4d4f18035", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-822446023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b28e6711f1b4b948dbfc124eca1e528", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45abb-82", "ovs_interfaceid": "e1a45abb-8278-47dc-9780-4592d326c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 814.611504] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:d7:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56398cc0-e39f-410f-8036-8c2a6870e26f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1a45abb-8278-47dc-9780-4592d326c13d', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.621500] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Creating folder: Project (8b28e6711f1b4b948dbfc124eca1e528). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.624289] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f9abeb6-e020-453e-b64f-be7b97fbab31 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.634698] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.635070] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.635264] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.635885] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.635885] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.635885] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.636197] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.636264] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.636431] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.636609] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.636789] env[63418]: DEBUG nova.virt.hardware [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.637683] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78db37e-2082-4f4b-9fe5-f686c9f424fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.642145] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Created folder: Project (8b28e6711f1b4b948dbfc124eca1e528) in parent group-v268354. [ 814.642341] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Creating folder: Instances. Parent ref: group-v268447. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.643019] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc27f94b-a27e-40e4-ad8a-d4b75a1a3097 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.649868] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b015a1-2409-4fe2-a1d7-566447bde066 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.655983] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Created folder: Instances in parent group-v268447. [ 814.656314] env[63418]: DEBUG oslo.service.loopingcall [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.657337] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.657969] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a64ae0c-2f29-4387-9f45-b3f2b8473bc2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.692671] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.692671] env[63418]: value = "task-1244964" [ 814.692671] env[63418]: _type = "Task" [ 814.692671] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.704141] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244964, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.807406] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526ecb96-5e33-a263-ea16-54dde4bd1cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.053455} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.807726] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.808009] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] fa4c0315-1ef5-4491-ab18-d49563b778fb/fa4c0315-1ef5-4491-ab18-d49563b778fb.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 814.808468] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62b33e2d-6a1a-4325-94a0-b313917cc357 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.817651] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 814.817651] env[63418]: value = "task-1244965" [ 814.817651] env[63418]: _type = "Task" [ 814.817651] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.827749] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.903262] env[63418]: INFO nova.compute.manager [-] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Took 1.41 seconds to deallocate network for instance. [ 814.912062] env[63418]: INFO nova.compute.manager [-] [instance: f632d71d-498f-4914-9895-8f37187a295f] Took 1.44 seconds to deallocate network for instance. [ 814.970388] env[63418]: DEBUG nova.compute.manager [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 814.972750] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6288125d-e6a9-4b3e-8d5a-ece7491b4562 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.986707] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244961, 'name': Destroy_Task, 'duration_secs': 0.809553} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.988327] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Destroyed the VM [ 814.988726] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 814.996909] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5d749027-8f25-4903-8a12-33c1c023a2a1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.002724] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 815.002724] env[63418]: value = "task-1244966" [ 815.002724] env[63418]: _type = "Task" [ 815.002724] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.011723] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244966, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.066100] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.070019] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 815.070256] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.774s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.070553] env[63418]: DEBUG nova.objects.instance [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lazy-loading 'resources' on Instance uuid 18327b62-d022-45e5-9ab0-71791491e61f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.190452] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.193667] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.205193] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244964, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.329330] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244965, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.414125] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.418395] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.496232] env[63418]: INFO nova.compute.manager [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] instance snapshotting [ 815.498438] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de579dc9-0c7a-4cbc-93f9-654b2009d039 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.522216] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee26f01-2448-4ca3-af50-88c43bc7a2e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.529208] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244966, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.574668] env[63418]: DEBUG nova.compute.utils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.584449] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 815.584555] env[63418]: DEBUG nova.network.neutron [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.656278] env[63418]: DEBUG nova.network.neutron [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Successfully updated port: f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.665691] env[63418]: DEBUG nova.policy [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6266794b1f04bf1b3ce28a935fe86ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07c684a54c5c47f39d479c8c8a415c79', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.693626] env[63418]: INFO nova.compute.manager [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Detaching volume b91a8528-5ded-45b1-b1bb-31b41d7fff4b [ 815.706605] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244964, 'name': CreateVM_Task, 'duration_secs': 0.600885} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.709210] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.709875] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.710053] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.710400] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.710674] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2c10103-197c-4daf-9802-2c64aab40067 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.715587] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 815.715587] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526c3ab6-3763-67cf-d2d4-d587bc4b0e24" [ 815.715587] env[63418]: _type = "Task" [ 815.715587] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.725877] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526c3ab6-3763-67cf-d2d4-d587bc4b0e24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.738914] env[63418]: INFO nova.virt.block_device [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Attempting to driver detach volume b91a8528-5ded-45b1-b1bb-31b41d7fff4b from mountpoint /dev/sdb [ 815.739181] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 815.739369] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268443', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'name': 'volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e7b9a60-9e4b-46e3-adf9-7337dc7d2171', 'attached_at': '', 'detached_at': '', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'serial': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 815.740614] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5e8129-7c17-4f3a-b72f-adc4b15faa7b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.767648] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73b63b4-3536-45ae-9dde-edffc434f9c4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.775511] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0947bcab-88a1-4da5-b3ff-012563084dd0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.799119] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb63053-831a-4364-8639-3045ab9ebdae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.816244] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] The volume has not been displaced from its original location: [datastore1] volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b/volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 815.821876] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Reconfiguring VM instance instance-00000033 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 815.825023] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-275ce4d2-a481-4a57-b66e-94a0759519ba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.847160] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525168} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.848527] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] fa4c0315-1ef5-4491-ab18-d49563b778fb/fa4c0315-1ef5-4491-ab18-d49563b778fb.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.848789] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.849142] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 815.849142] env[63418]: value = "task-1244967" [ 815.849142] env[63418]: _type = "Task" [ 815.849142] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.851655] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a85cf63e-26f8-4eeb-b431-743999718852 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.863372] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244967, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.868012] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 815.868012] env[63418]: value = "task-1244968" [ 815.868012] env[63418]: _type = "Task" [ 815.868012] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.877511] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.962942] env[63418]: DEBUG nova.network.neutron [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Successfully created port: 43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.011885] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Received event network-changed-e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 816.011885] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Refreshing instance network info cache due to event network-changed-e1a45abb-8278-47dc-9780-4592d326c13d. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 816.011885] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquiring lock "refresh_cache-f7554ae4-c7a7-4111-a830-10f9029dc074" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.012524] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquired lock "refresh_cache-f7554ae4-c7a7-4111-a830-10f9029dc074" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.012748] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Refreshing network info cache for port e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.021293] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fd66f4-a4ee-4a7d-bf33-dd5d440aad59 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.022926] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244966, 'name': RemoveSnapshot_Task, 'duration_secs': 0.630071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.023704] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 816.024033] env[63418]: DEBUG nova.compute.manager [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 816.025307] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3126de52-fb5d-4b81-845c-e0e92c4f3949 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.032358] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d366db-a0f5-4fce-aea6-8caded0aff08 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.041257] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 816.044311] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0d44d514-eb52-4827-944b-d0298aeda109 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.084987] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 816.092029] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a64bd1-be27-4751-90d4-5b03933011df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.093698] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 816.093698] env[63418]: value = "task-1244969" [ 816.093698] env[63418]: _type = "Task" [ 816.093698] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.102347] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b62439-228c-4eaa-8509-bf4e881dcfc9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.112229] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244969, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.124464] env[63418]: DEBUG nova.compute.provider_tree [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 816.168082] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.168249] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.168407] env[63418]: DEBUG nova.network.neutron [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.226915] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526c3ab6-3763-67cf-d2d4-d587bc4b0e24, 'name': SearchDatastore_Task, 'duration_secs': 0.008613} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.227248] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.227487] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.227727] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.227905] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.228200] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.228493] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5128b504-f0b1-4dab-8f6e-44ca96e68e3b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.236960] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.241024] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.241024] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3caffe8-eb68-4bce-b4e7-d276c56504b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.243653] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 816.243653] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528eca2f-bdd2-3506-d3f9-5d9f1600ef71" [ 816.243653] env[63418]: _type = "Task" [ 816.243653] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.251661] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528eca2f-bdd2-3506-d3f9-5d9f1600ef71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.365375] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.377644] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07118} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.377731] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.378565] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75a6e88-8d3a-4b36-8089-6a1b781788ff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.401028] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] fa4c0315-1ef5-4491-ab18-d49563b778fb/fa4c0315-1ef5-4491-ab18-d49563b778fb.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.401238] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ad7db38-e4ff-446b-80c6-8ee79d5c6f07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.423039] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 816.423039] env[63418]: value = "task-1244970" [ 816.423039] env[63418]: _type = "Task" [ 816.423039] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.432042] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.555072] env[63418]: INFO nova.compute.manager [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Shelve offloading [ 816.609467] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244969, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.617769] env[63418]: DEBUG nova.compute.manager [req-9744f582-3cbf-43f1-8c5e-6290263e3fe5 req-97f7077f-08de-4a0a-98e9-cbce6e09a665 service nova] [instance: f632d71d-498f-4914-9895-8f37187a295f] Received event network-vif-deleted-02aff9ac-be86-40b7-8b63-a237fd928606 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 816.676331] env[63418]: DEBUG nova.scheduler.client.report [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 79 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 816.676331] env[63418]: DEBUG nova.compute.provider_tree [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 79 to 80 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 816.676914] env[63418]: DEBUG nova.compute.provider_tree [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 816.717156] env[63418]: DEBUG nova.network.neutron [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.757467] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528eca2f-bdd2-3506-d3f9-5d9f1600ef71, 'name': SearchDatastore_Task, 'duration_secs': 0.038637} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.760988] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-511ec2d5-503b-44cd-847a-45945765d450 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.767155] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 816.767155] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52140547-759f-b066-3e75-a7800edd51e3" [ 816.767155] env[63418]: _type = "Task" [ 816.767155] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.777014] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52140547-759f-b066-3e75-a7800edd51e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.865038] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244967, 'name': ReconfigVM_Task, 'duration_secs': 0.676741} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.865439] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Reconfigured VM instance instance-00000033 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 816.870420] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6a140a7-48ef-41dc-a84d-b8aa385dce15 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.881935] env[63418]: DEBUG nova.network.neutron [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.889054] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 816.889054] env[63418]: value = "task-1244971" [ 816.889054] env[63418]: _type = "Task" [ 816.889054] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.897687] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.935586] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244970, 'name': ReconfigVM_Task, 'duration_secs': 0.291252} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.935882] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Reconfigured VM instance instance-00000041 to attach disk [datastore1] fa4c0315-1ef5-4491-ab18-d49563b778fb/fa4c0315-1ef5-4491-ab18-d49563b778fb.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.936615] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa5158db-ff16-46b9-8b72-adb632b5d82e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.944373] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Updated VIF entry in instance network info cache for port e1a45abb-8278-47dc-9780-4592d326c13d. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 816.944731] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Updating instance_info_cache with network_info: [{"id": "e1a45abb-8278-47dc-9780-4592d326c13d", "address": "fa:16:3e:c5:d7:f5", "network": {"id": "67e9ab9a-8452-4b02-b84c-1fa4d4f18035", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-822446023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b28e6711f1b4b948dbfc124eca1e528", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45abb-82", "ovs_interfaceid": "e1a45abb-8278-47dc-9780-4592d326c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.947598] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 816.947598] env[63418]: value = "task-1244972" [ 816.947598] env[63418]: _type = "Task" [ 816.947598] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.956609] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244972, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.062028] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.062028] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-147ecdd0-70f2-4019-9395-828e9a13048a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.069755] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 817.069755] env[63418]: value = "task-1244973" [ 817.069755] env[63418]: _type = "Task" [ 817.069755] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.078473] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244973, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.107696] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 817.115963] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244969, 'name': CreateSnapshot_Task, 'duration_secs': 0.946496} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.116413] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 817.117099] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b4f9a3-3b8b-4498-84fb-370df9635d56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.136132] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.136425] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.136682] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.136765] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.136911] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.137076] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.137293] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.137455] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.137625] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.137791] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.138016] env[63418]: DEBUG nova.virt.hardware [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.138886] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1ac4b8-5642-436c-9df8-94f128a9a9c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.148978] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c2c789-c887-4f08-8d2d-dd00295bdb2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.183578] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.113s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.187401] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.149s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.187628] env[63418]: DEBUG nova.objects.instance [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lazy-loading 'resources' on Instance uuid 17a055e4-13da-4984-ae64-af1ae9d652f6 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.216588] env[63418]: INFO nova.scheduler.client.report [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Deleted allocations for instance 18327b62-d022-45e5-9ab0-71791491e61f [ 817.279018] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52140547-759f-b066-3e75-a7800edd51e3, 'name': SearchDatastore_Task, 'duration_secs': 0.012746} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.279684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.279684] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] f7554ae4-c7a7-4111-a830-10f9029dc074/f7554ae4-c7a7-4111-a830-10f9029dc074.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.279876] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c96805e5-6bf6-4564-8190-8d6555b3a8d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.286591] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 817.286591] env[63418]: value = "task-1244974" [ 817.286591] env[63418]: _type = "Task" [ 817.286591] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.295170] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244974, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.385243] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.385637] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance network_info: |[{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 817.386162] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:c4:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f786c95b-5214-454d-86f8-6d922f0482d8', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.395320] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating folder: Project (0462b212fa4449c2a6f98cec2f186f51). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.395735] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4e4ef5f-8861-46e4-8101-4f7d1a09fdeb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.407265] env[63418]: DEBUG oslo_vmware.api [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1244971, 'name': ReconfigVM_Task, 'duration_secs': 0.168103} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.407669] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268443', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'name': 'volume-b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e7b9a60-9e4b-46e3-adf9-7337dc7d2171', 'attached_at': '', 'detached_at': '', 'volume_id': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b', 'serial': 'b91a8528-5ded-45b1-b1bb-31b41d7fff4b'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 817.411626] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Created folder: Project (0462b212fa4449c2a6f98cec2f186f51) in parent group-v268354. [ 817.411626] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating folder: Instances. Parent ref: group-v268451. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.412077] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62788d4f-aae4-481b-921e-406e75f3f695 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.420640] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Created folder: Instances in parent group-v268451. [ 817.420936] env[63418]: DEBUG oslo.service.loopingcall [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.421200] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.421455] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6e5e8a6-12b4-440f-a4fc-8845eb0c8208 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.444083] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.444083] env[63418]: value = "task-1244977" [ 817.444083] env[63418]: _type = "Task" [ 817.444083] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.449612] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Releasing lock "refresh_cache-f7554ae4-c7a7-4111-a830-10f9029dc074" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.449997] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-vif-plugged-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 817.454101] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.454101] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.454101] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.454101] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] No waiting events found dispatching network-vif-plugged-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 817.454101] env[63418]: WARNING nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received unexpected event network-vif-plugged-f786c95b-5214-454d-86f8-6d922f0482d8 for instance with vm_state building and task_state spawning. [ 817.454282] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-changed-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 817.454282] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Refreshing instance network info cache due to event network-changed-f786c95b-5214-454d-86f8-6d922f0482d8. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 817.454282] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.454282] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.454282] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Refreshing network info cache for port f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.463414] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244977, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.470333] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244972, 'name': Rename_Task, 'duration_secs': 0.179437} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.470658] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.470972] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03bb9ce1-17c7-46ef-b8f8-4e5c23177250 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.477731] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 817.477731] env[63418]: value = "task-1244978" [ 817.477731] env[63418]: _type = "Task" [ 817.477731] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.487828] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244978, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.532850] env[63418]: DEBUG nova.network.neutron [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Successfully updated port: 43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.582604] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 817.582867] env[63418]: DEBUG nova.compute.manager [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 817.583684] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee643f4-3427-464b-9a49-e300591a164d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.591166] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.591423] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.591711] env[63418]: DEBUG nova.network.neutron [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.637902] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 817.638870] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-43be6ee6-f100-48f5-bfef-98f3902d9932 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.650096] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 817.650096] env[63418]: value = "task-1244979" [ 817.650096] env[63418]: _type = "Task" [ 817.650096] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.662924] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244979, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.726522] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7f583ec-2994-4239-8b65-8fd3f04e5a27 tempest-InstanceActionsV221TestJSON-231865196 tempest-InstanceActionsV221TestJSON-231865196-project-member] Lock "18327b62-d022-45e5-9ab0-71791491e61f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.401s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.799745] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244974, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.954199] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244977, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.970333] env[63418]: DEBUG nova.objects.instance [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lazy-loading 'flavor' on Instance uuid 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.993575] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244978, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.035978] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "refresh_cache-849552d4-0bb4-48ee-af7b-390183bd2189" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.036154] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquired lock "refresh_cache-849552d4-0bb4-48ee-af7b-390183bd2189" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.036305] env[63418]: DEBUG nova.network.neutron [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.047520] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c95fb6b-1268-42ed-b25a-3610d40fb289 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.056852] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa75bbb-2778-436c-b5bc-d61ab6ea2db3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.095789] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ea6bdb-2517-43f7-95e6-7cfe8a971311 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.106954] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327a8360-3363-47d9-a901-d3182d0d515f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.121919] env[63418]: DEBUG nova.compute.provider_tree [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.162713] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244979, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.215387] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updated VIF entry in instance network info cache for port f786c95b-5214-454d-86f8-6d922f0482d8. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.215782] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.303722] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244974, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573663} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.304258] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] f7554ae4-c7a7-4111-a830-10f9029dc074/f7554ae4-c7a7-4111-a830-10f9029dc074.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.305317] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.305750] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1c22e0c-91b4-44dc-9e48-f1d9bdb95655 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.316869] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 818.316869] env[63418]: value = "task-1244980" [ 818.316869] env[63418]: _type = "Task" [ 818.316869] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.329788] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244980, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.455966] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244977, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.488589] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244978, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.571623] env[63418]: DEBUG nova.network.neutron [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.627777] env[63418]: DEBUG nova.scheduler.client.report [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 818.662410] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244979, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.717020] env[63418]: DEBUG nova.network.neutron [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.718489] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.718841] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Received event network-changed-d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 818.718912] env[63418]: DEBUG nova.compute.manager [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Refreshing instance network info cache due to event network-changed-d0fa119b-63f3-4b54-8592-195a043ff0ee. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 818.719109] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.719399] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.719456] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Refreshing network info cache for port d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.746524] env[63418]: DEBUG nova.network.neutron [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Updating instance_info_cache with network_info: [{"id": "43c32def-abd1-4085-95c5-a30bfcbcf4e0", "address": "fa:16:3e:45:03:a6", "network": {"id": "4a8c77d7-ee94-4dc6-9b18-517345b55eb6", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-396982631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07c684a54c5c47f39d479c8c8a415c79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c32def-ab", "ovs_interfaceid": "43c32def-abd1-4085-95c5-a30bfcbcf4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.797651] env[63418]: DEBUG nova.compute.manager [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Received event network-vif-plugged-43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 818.797869] env[63418]: DEBUG oslo_concurrency.lockutils [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] Acquiring lock "849552d4-0bb4-48ee-af7b-390183bd2189-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.798117] env[63418]: DEBUG oslo_concurrency.lockutils [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] Lock "849552d4-0bb4-48ee-af7b-390183bd2189-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.798240] env[63418]: DEBUG oslo_concurrency.lockutils [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] Lock "849552d4-0bb4-48ee-af7b-390183bd2189-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.798436] env[63418]: DEBUG nova.compute.manager [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] No waiting events found dispatching network-vif-plugged-43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 818.798610] env[63418]: WARNING nova.compute.manager [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Received unexpected event network-vif-plugged-43c32def-abd1-4085-95c5-a30bfcbcf4e0 for instance with vm_state building and task_state spawning. [ 818.798828] env[63418]: DEBUG nova.compute.manager [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Received event network-changed-43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 818.798945] env[63418]: DEBUG nova.compute.manager [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Refreshing instance network info cache due to event network-changed-43c32def-abd1-4085-95c5-a30bfcbcf4e0. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 818.799133] env[63418]: DEBUG oslo_concurrency.lockutils [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] Acquiring lock "refresh_cache-849552d4-0bb4-48ee-af7b-390183bd2189" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.826473] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244980, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08768} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.827106] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.827917] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245ec30f-6e83-4b46-ac6c-f7abf83cb7e7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.850728] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] f7554ae4-c7a7-4111-a830-10f9029dc074/f7554ae4-c7a7-4111-a830-10f9029dc074.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.850728] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ba2ed70-94f6-4862-be7b-904f9b5ed8b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.870378] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 818.870378] env[63418]: value = "task-1244981" [ 818.870378] env[63418]: _type = "Task" [ 818.870378] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.878824] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.955220] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244977, 'name': CreateVM_Task, 'duration_secs': 1.050197} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.955406] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 818.956105] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.956279] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.956617] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 818.956870] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c7bde9d-e5aa-4b14-85c6-d358fa889d9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.961502] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 818.961502] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5231a076-259c-5402-fce0-8051411cada5" [ 818.961502] env[63418]: _type = "Task" [ 818.961502] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.970059] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5231a076-259c-5402-fce0-8051411cada5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.979280] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c0b0035e-bc8b-4e20-81fb-2b1471d117c5 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.788s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.988469] env[63418]: DEBUG oslo_vmware.api [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1244978, 'name': PowerOnVM_Task, 'duration_secs': 1.177608} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.988749] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.988912] env[63418]: INFO nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Took 12.40 seconds to spawn the instance on the hypervisor. [ 818.989109] env[63418]: DEBUG nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 818.991480] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6428d3d9-f1f7-4c09-a380-f41ce53902db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.002832] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248454d-4c73-737a-613a-bf9421151849/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 819.003852] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f75938-a7f8-4f70-9a67-f0cf5ab92dc0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.009575] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248454d-4c73-737a-613a-bf9421151849/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 819.009741] env[63418]: ERROR oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248454d-4c73-737a-613a-bf9421151849/disk-0.vmdk due to incomplete transfer. [ 819.009950] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b6fe2a35-a065-476d-a751-11880061c5dc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.016699] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248454d-4c73-737a-613a-bf9421151849/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 819.016940] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Uploaded image 73126df0-d2dd-40c7-9e81-e5d195e1d214 to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 819.020131] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 819.020131] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2c629faf-1f90-401a-a438-0d00e61264bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.024662] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 819.024662] env[63418]: value = "task-1244982" [ 819.024662] env[63418]: _type = "Task" [ 819.024662] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.031799] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244982, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.133026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.133026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.560s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.133243] env[63418]: DEBUG nova.objects.instance [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 819.157056] env[63418]: INFO nova.scheduler.client.report [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleted allocations for instance 17a055e4-13da-4984-ae64-af1ae9d652f6 [ 819.166095] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244979, 'name': CloneVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.221593] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.252022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Releasing lock "refresh_cache-849552d4-0bb4-48ee-af7b-390183bd2189" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.252022] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Instance network_info: |[{"id": "43c32def-abd1-4085-95c5-a30bfcbcf4e0", "address": "fa:16:3e:45:03:a6", "network": {"id": "4a8c77d7-ee94-4dc6-9b18-517345b55eb6", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-396982631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07c684a54c5c47f39d479c8c8a415c79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c32def-ab", "ovs_interfaceid": "43c32def-abd1-4085-95c5-a30bfcbcf4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 819.252296] env[63418]: DEBUG oslo_concurrency.lockutils [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] Acquired lock "refresh_cache-849552d4-0bb4-48ee-af7b-390183bd2189" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.252296] env[63418]: DEBUG nova.network.neutron [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Refreshing network info cache for port 43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.252659] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:03:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9875d38f-76e2-416c-bfb7-f18a22b0d8ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43c32def-abd1-4085-95c5-a30bfcbcf4e0', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.261328] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Creating folder: Project (07c684a54c5c47f39d479c8c8a415c79). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.265136] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d54b82ed-e414-4578-afd6-f3d2fb577082 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.282953] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Created folder: Project (07c684a54c5c47f39d479c8c8a415c79) in parent group-v268354. [ 819.282953] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Creating folder: Instances. Parent ref: group-v268455. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.282953] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f08eb121-1ef7-4d3d-8341-b6923feb3665 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.296250] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Created folder: Instances in parent group-v268455. [ 819.296250] env[63418]: DEBUG oslo.service.loopingcall [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.296250] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 819.296250] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-827e769f-040e-4d36-bb93-837ec5da5258 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.317305] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.317305] env[63418]: value = "task-1244985" [ 819.317305] env[63418]: _type = "Task" [ 819.317305] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.330989] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244985, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.388548] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244981, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.473189] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5231a076-259c-5402-fce0-8051411cada5, 'name': SearchDatastore_Task, 'duration_secs': 0.032622} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.474111] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.474534] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.475108] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.475719] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.476182] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.477104] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c14d3824-39eb-4b91-ab67-0bf70ce30518 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.489152] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.489152] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.489152] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c11f594-1028-4ef3-a655-1f6f1bd9c7a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.496454] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 819.496454] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528a26f2-ae0f-e710-36f2-62491d506641" [ 819.496454] env[63418]: _type = "Task" [ 819.496454] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.505707] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528a26f2-ae0f-e710-36f2-62491d506641, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.517851] env[63418]: INFO nova.compute.manager [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Took 34.04 seconds to build instance. [ 819.534224] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244982, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.662755] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244979, 'name': CloneVM_Task, 'duration_secs': 1.76622} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.663531] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Created linked-clone VM from snapshot [ 819.665010] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f599b00d-f6be-4553-8d64-6c0298f31a58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.675964] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Uploading image 531bb859-dc73-4a87-86b7-2978bf9f4e60 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 819.678423] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74bc32ad-046a-4d6a-9da7-c977201502ac tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "17a055e4-13da-4984-ae64-af1ae9d652f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.108s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.693571] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updated VIF entry in instance network info cache for port d0fa119b-63f3-4b54-8592-195a043ff0ee. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.693571] env[63418]: DEBUG nova.network.neutron [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.696590] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 819.696829] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-93f99618-08b0-4cc0-b00a-1ade6a53885a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.704089] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 819.704089] env[63418]: value = "task-1244986" [ 819.704089] env[63418]: _type = "Task" [ 819.704089] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.714981] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244986, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.829580] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244985, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.880480] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244981, 'name': ReconfigVM_Task, 'duration_secs': 0.54074} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.880971] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Reconfigured VM instance instance-00000042 to attach disk [datastore1] f7554ae4-c7a7-4111-a830-10f9029dc074/f7554ae4-c7a7-4111-a830-10f9029dc074.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.881694] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f7e48ce-47e1-4bdc-ad8a-f8e7dc4ae051 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.888530] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 819.888530] env[63418]: value = "task-1244987" [ 819.888530] env[63418]: _type = "Task" [ 819.888530] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.897431] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244987, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.007064] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528a26f2-ae0f-e710-36f2-62491d506641, 'name': SearchDatastore_Task, 'duration_secs': 0.023147} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.007895] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fb3aec0-93af-49b0-98ad-4dcb264864eb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.013841] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 820.013841] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525ca475-b593-75bd-4731-cbb220004471" [ 820.013841] env[63418]: _type = "Task" [ 820.013841] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.022560] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3886f622-1dfc-41db-91c0-33329516c124 tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.096s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.022880] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525ca475-b593-75bd-4731-cbb220004471, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.039606] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244982, 'name': Destroy_Task, 'duration_secs': 0.830452} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.039770] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Destroyed the VM [ 820.040034] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 820.040299] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5ae2c847-8446-4458-a5a6-8d29f991d6cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.047520] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 820.047520] env[63418]: value = "task-1244988" [ 820.047520] env[63418]: _type = "Task" [ 820.047520] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.057281] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244988, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.096220] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.097166] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ea45ff-0826-44c9-a5fd-08013f3f7cff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.107479] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.107757] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6caa255d-a24b-4f82-b5c9-20c958d57979 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.144472] env[63418]: DEBUG oslo_concurrency.lockutils [None req-612865f4-61a8-4218-9a77-a0aa3808f649 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.145683] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.068s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.149207] env[63418]: INFO nova.compute.claims [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.166604] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.167046] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.167046] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleting the datastore file [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.167310] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf7b02e9-dd78-4de8-8d47-e824945a052d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.178071] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 820.178071] env[63418]: value = "task-1244990" [ 820.178071] env[63418]: _type = "Task" [ 820.178071] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.189823] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.195646] env[63418]: DEBUG oslo_concurrency.lockutils [req-f4aa709b-3f33-418c-913a-1f92a81457f1 req-3d3cadb2-1716-4ca8-9070-c3f7e2a01224 service nova] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.214183] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244986, 'name': Destroy_Task, 'duration_secs': 0.37676} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.215254] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Destroyed the VM [ 820.215511] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 820.215772] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4b20e20a-4fc2-4637-9d56-4859d5f26b71 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.224340] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 820.224340] env[63418]: value = "task-1244991" [ 820.224340] env[63418]: _type = "Task" [ 820.224340] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.231818] env[63418]: DEBUG nova.network.neutron [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Updated VIF entry in instance network info cache for port 43c32def-abd1-4085-95c5-a30bfcbcf4e0. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.232212] env[63418]: DEBUG nova.network.neutron [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Updating instance_info_cache with network_info: [{"id": "43c32def-abd1-4085-95c5-a30bfcbcf4e0", "address": "fa:16:3e:45:03:a6", "network": {"id": "4a8c77d7-ee94-4dc6-9b18-517345b55eb6", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-396982631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07c684a54c5c47f39d479c8c8a415c79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c32def-ab", "ovs_interfaceid": "43c32def-abd1-4085-95c5-a30bfcbcf4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.238065] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244991, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.333554] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244985, 'name': CreateVM_Task, 'duration_secs': 0.577287} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.333730] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 820.334524] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.334684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.335046] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.335313] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-453ce2f4-bf4f-47b9-b9e6-f17c45facca3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.345420] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 820.345420] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e9b4d9-be08-91cb-d8e3-39c5bdefc8e3" [ 820.345420] env[63418]: _type = "Task" [ 820.345420] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.355064] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e9b4d9-be08-91cb-d8e3-39c5bdefc8e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.402623] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244987, 'name': Rename_Task, 'duration_secs': 0.149411} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.402865] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.403140] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff874076-d3eb-4ee8-9aec-beb1f28d24c9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.411865] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 820.411865] env[63418]: value = "task-1244992" [ 820.411865] env[63418]: _type = "Task" [ 820.411865] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.423194] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.532855] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525ca475-b593-75bd-4731-cbb220004471, 'name': SearchDatastore_Task, 'duration_secs': 0.021136} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.533176] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.533446] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.533724] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d78835f-3d2a-4c87-a5ac-789f75770f5d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.539966] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 820.539966] env[63418]: value = "task-1244993" [ 820.539966] env[63418]: _type = "Task" [ 820.539966] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.549999] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.559150] env[63418]: DEBUG oslo_vmware.api [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1244988, 'name': RemoveSnapshot_Task, 'duration_secs': 0.496983} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.559412] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 820.559643] env[63418]: INFO nova.compute.manager [None req-2db7694e-37cf-466b-97b8-83320aed4b24 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Took 18.69 seconds to snapshot the instance on the hypervisor. [ 820.685422] env[63418]: DEBUG oslo_vmware.api [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1244990, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158092} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.685634] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.685850] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.685963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.720213] env[63418]: INFO nova.scheduler.client.report [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted allocations for instance d76a008c-9bd9-420b-873d-4f7d7f25b8ca [ 820.739816] env[63418]: DEBUG oslo_concurrency.lockutils [req-8fee6a9e-2e81-4fe7-9146-81f6b8ee005d req-b171c386-2336-4569-bd48-c9fbc7f1237b service nova] Releasing lock "refresh_cache-849552d4-0bb4-48ee-af7b-390183bd2189" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.740287] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244991, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.786148] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "6b81dbe2-aa9e-4561-962b-2af167234b90" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.786518] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.786752] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "6b81dbe2-aa9e-4561-962b-2af167234b90-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.786951] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.787144] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.790048] env[63418]: INFO nova.compute.manager [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Terminating instance [ 820.858673] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e9b4d9-be08-91cb-d8e3-39c5bdefc8e3, 'name': SearchDatastore_Task, 'duration_secs': 0.009497} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.861868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.861868] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.861868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.861868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.862118] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.862118] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ace89440-87d4-4648-8790-54864ce832fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.871427] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.871667] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.872673] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0631bd9-7e03-42ea-b91d-969a17db6ece {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.877924] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 820.877924] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527c578e-1b3d-9dc2-3a0b-75e8e2afb43c" [ 820.877924] env[63418]: _type = "Task" [ 820.877924] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.887774] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527c578e-1b3d-9dc2-3a0b-75e8e2afb43c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.924510] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244992, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.928741] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "e19a60cd-4ec7-48ad-9042-f19b75353364" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.928987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.050738] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.118032] env[63418]: DEBUG nova.compute.manager [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-vif-unplugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 821.118316] env[63418]: DEBUG oslo_concurrency.lockutils [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.118530] env[63418]: DEBUG oslo_concurrency.lockutils [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.118645] env[63418]: DEBUG oslo_concurrency.lockutils [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.119116] env[63418]: DEBUG nova.compute.manager [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] No waiting events found dispatching network-vif-unplugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 821.119282] env[63418]: WARNING nova.compute.manager [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received unexpected event network-vif-unplugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 for instance with vm_state shelved_offloaded and task_state None. [ 821.119482] env[63418]: DEBUG nova.compute.manager [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 821.119664] env[63418]: DEBUG nova.compute.manager [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing instance network info cache due to event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 821.119887] env[63418]: DEBUG oslo_concurrency.lockutils [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.120066] env[63418]: DEBUG oslo_concurrency.lockutils [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.120256] env[63418]: DEBUG nova.network.neutron [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.226254] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.240498] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244991, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.297325] env[63418]: DEBUG nova.compute.manager [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 821.297583] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.302839] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7098686-ad02-4a67-b95a-4eb0b3edcd85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.313823] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.314250] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ef0c232-cb31-4b56-9837-7038e0013c04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.326258] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 821.326258] env[63418]: value = "task-1244994" [ 821.326258] env[63418]: _type = "Task" [ 821.326258] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.336474] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.389355] env[63418]: DEBUG nova.compute.manager [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Received event network-changed-2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 821.389617] env[63418]: DEBUG nova.compute.manager [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Refreshing instance network info cache due to event network-changed-2024d711-87c3-497d-8373-842a53f06798. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 821.389876] env[63418]: DEBUG oslo_concurrency.lockutils [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] Acquiring lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.390094] env[63418]: DEBUG oslo_concurrency.lockutils [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] Acquired lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.390281] env[63418]: DEBUG nova.network.neutron [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Refreshing network info cache for port 2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.397523] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527c578e-1b3d-9dc2-3a0b-75e8e2afb43c, 'name': SearchDatastore_Task, 'duration_secs': 0.008708} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.403399] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43197f7f-2267-44b2-b8a4-2ba2101bc61b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.413731] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 821.413731] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]524866be-ce51-a96e-9fdf-6465a5bef5c5" [ 821.413731] env[63418]: _type = "Task" [ 821.413731] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.431025] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524866be-ce51-a96e-9fdf-6465a5bef5c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.431471] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 821.438992] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244992, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.532069] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bbe206-c619-423c-88bd-784ebb89a225 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.554508] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73889fe4-bf29-4b16-895a-37889f56be12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.563344] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.977946} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.564136] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.564406] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.564691] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-833633a6-b645-49fa-a8a5-5178a86296fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.594466] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4ef47c-6302-4789-b88a-708c9ffb0d0e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.598716] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 821.598716] env[63418]: value = "task-1244995" [ 821.598716] env[63418]: _type = "Task" [ 821.598716] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.605926] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aac27ae-0442-4286-866a-4bc2150af3a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.614137] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244995, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.627996] env[63418]: DEBUG nova.compute.provider_tree [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.737778] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244991, 'name': RemoveSnapshot_Task} progress is 76%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.763673] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.764147] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.764480] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.764780] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.765086] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.768288] env[63418]: INFO nova.compute.manager [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Terminating instance [ 821.835665] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244994, 'name': PowerOffVM_Task, 'duration_secs': 0.278424} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.836070] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 821.836295] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 821.836563] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ac0e778-de35-408b-8b83-1373e88cf559 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.918114] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 821.918390] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 821.918390] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleting the datastore file [datastore2] 6b81dbe2-aa9e-4561-962b-2af167234b90 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.919119] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-264e6c69-aa7f-499d-a308-712de90adbd3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.929245] env[63418]: DEBUG oslo_vmware.api [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1244992, 'name': PowerOnVM_Task, 'duration_secs': 1.487073} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.934034] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.934034] env[63418]: INFO nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Took 9.84 seconds to spawn the instance on the hypervisor. [ 821.934034] env[63418]: DEBUG nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 821.934284] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524866be-ce51-a96e-9fdf-6465a5bef5c5, 'name': SearchDatastore_Task, 'duration_secs': 0.029874} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.935741] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for the task: (returnval){ [ 821.935741] env[63418]: value = "task-1244997" [ 821.935741] env[63418]: _type = "Task" [ 821.935741] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.935741] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c1c403-1feb-4557-b16c-bf6a09e0d4fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.937677] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.938024] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 849552d4-0bb4-48ee-af7b-390183bd2189/849552d4-0bb4-48ee-af7b-390183bd2189.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.945507] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50ae59de-9e89-4848-8b77-5789a894c58b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.960884] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 821.960884] env[63418]: value = "task-1244998" [ 821.960884] env[63418]: _type = "Task" [ 821.960884] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.961671] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244997, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.966639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.972334] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1244998, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.039017] env[63418]: DEBUG nova.network.neutron [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updated VIF entry in instance network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.039017] env[63418]: DEBUG nova.network.neutron [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0263ccbe-85", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.045586] env[63418]: DEBUG oslo_concurrency.lockutils [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.045586] env[63418]: DEBUG oslo_concurrency.lockutils [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.045586] env[63418]: INFO nova.compute.manager [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Rebooting instance [ 822.109165] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244995, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073919} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.110102] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.111664] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bc59a4-46a0-4bb1-8d59-6bb606507036 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.131155] env[63418]: DEBUG nova.scheduler.client.report [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 822.146758] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.146758] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85e0ab48-59fd-4a93-9464-4a379cdc1c2d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.162448] env[63418]: DEBUG nova.network.neutron [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updated VIF entry in instance network info cache for port 2024d711-87c3-497d-8373-842a53f06798. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.162815] env[63418]: DEBUG nova.network.neutron [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [{"id": "2024d711-87c3-497d-8373-842a53f06798", "address": "fa:16:3e:e9:da:bf", "network": {"id": "3c73f0ee-1a86-4875-8c44-5658ef2ff43f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1851341556-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c828658cd364400a9207d0c90f3ae1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2024d711-87", "ovs_interfaceid": "2024d711-87c3-497d-8373-842a53f06798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.170941] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 822.170941] env[63418]: value = "task-1244999" [ 822.170941] env[63418]: _type = "Task" [ 822.170941] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.180710] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244999, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.236525] env[63418]: DEBUG oslo_vmware.api [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1244991, 'name': RemoveSnapshot_Task, 'duration_secs': 1.798183} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.236750] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 822.272924] env[63418]: DEBUG nova.compute.manager [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 822.273226] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.274510] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbb32cf-e384-48e3-8dbc-0ce49ac696df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.282077] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.282333] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-612d6fdc-3aeb-4561-a4bd-1dc1479b6765 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.351056] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.351056] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.351227] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleting the datastore file [datastore2] 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.351708] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f3a35eb-039c-444f-bc61-c2e1408ad20f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.361024] env[63418]: DEBUG oslo_vmware.api [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 822.361024] env[63418]: value = "task-1245001" [ 822.361024] env[63418]: _type = "Task" [ 822.361024] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.371779] env[63418]: DEBUG oslo_vmware.api [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.449445] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244997, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.474115] env[63418]: INFO nova.compute.manager [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Took 31.73 seconds to build instance. [ 822.478777] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1244998, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.541784] env[63418]: DEBUG oslo_concurrency.lockutils [req-f96118eb-5a9d-4253-815b-da42f3fbc42e req-df4e474e-d40f-4b86-9ff0-f2761ae60c3a service nova] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.573142] env[63418]: DEBUG oslo_concurrency.lockutils [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.635139] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.635462] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.641019] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.641019] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 822.642620] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.790s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.642846] env[63418]: DEBUG nova.objects.instance [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lazy-loading 'resources' on Instance uuid 27290558-6c58-414d-bd53-c4a686890721 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.667620] env[63418]: DEBUG oslo_concurrency.lockutils [req-8078b46e-d5d3-4514-b2f7-b90ba0cd0033 req-0b7259f8-a1aa-4180-8ff0-168ab11c8e18 service nova] Releasing lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.667620] env[63418]: DEBUG oslo_concurrency.lockutils [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquired lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.667620] env[63418]: DEBUG nova.network.neutron [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.684497] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.742879] env[63418]: WARNING nova.compute.manager [None req-a78c3706-9c52-4e29-b746-7f7f83d923ec tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Image not found during snapshot: nova.exception.ImageNotFound: Image 531bb859-dc73-4a87-86b7-2978bf9f4e60 could not be found. [ 822.870325] env[63418]: DEBUG oslo_vmware.api [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.948883] env[63418]: DEBUG oslo_vmware.api [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Task: {'id': task-1244997, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.530527} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.949206] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.949397] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 822.949575] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.949749] env[63418]: INFO nova.compute.manager [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Took 1.65 seconds to destroy the instance on the hypervisor. [ 822.950126] env[63418]: DEBUG oslo.service.loopingcall [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.951021] env[63418]: DEBUG nova.compute.manager [-] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 822.951021] env[63418]: DEBUG nova.network.neutron [-] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.976071] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1244998, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.923275} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.976255] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 849552d4-0bb4-48ee-af7b-390183bd2189/849552d4-0bb4-48ee-af7b-390183bd2189.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.976487] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.977046] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2574539-820c-4bc4-99a2-87503ec13d25 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.980137] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e4b0f148-c753-4141-bc39-1b6ff75d472e tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.004s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.983674] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 822.983674] env[63418]: value = "task-1245002" [ 822.983674] env[63418]: _type = "Task" [ 822.983674] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.992280] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.112202] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "f7554ae4-c7a7-4111-a830-10f9029dc074" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.112202] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.112594] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "f7554ae4-c7a7-4111-a830-10f9029dc074-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.112594] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.112816] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.116823] env[63418]: INFO nova.compute.manager [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Terminating instance [ 823.140069] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 823.143762] env[63418]: DEBUG nova.compute.utils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.145682] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 823.145935] env[63418]: DEBUG nova.network.neutron [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.185154] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1244999, 'name': ReconfigVM_Task, 'duration_secs': 0.609836} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.186682] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.189027] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3136d7d-14a4-4618-a33c-ac74b13d3627 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.204020] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 823.204020] env[63418]: value = "task-1245003" [ 823.204020] env[63418]: _type = "Task" [ 823.204020] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.211273] env[63418]: DEBUG nova.policy [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ff64612530b451fb41100a5aa601be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d52f1fceb24234a8d967038b43c857', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.223592] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245003, 'name': Rename_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.362334] env[63418]: DEBUG nova.compute.manager [req-23f28a90-b6fa-4d88-a936-d4eddd2c632e req-ae9a8f46-d1b7-42f4-9644-8aeac3a7fd0f service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Received event network-vif-deleted-9d15beed-d1a9-4234-b1ab-fd8df3c8abc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 823.362539] env[63418]: INFO nova.compute.manager [req-23f28a90-b6fa-4d88-a936-d4eddd2c632e req-ae9a8f46-d1b7-42f4-9644-8aeac3a7fd0f service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Neutron deleted interface 9d15beed-d1a9-4234-b1ab-fd8df3c8abc6; detaching it from the instance and deleting it from the info cache [ 823.362711] env[63418]: DEBUG nova.network.neutron [req-23f28a90-b6fa-4d88-a936-d4eddd2c632e req-ae9a8f46-d1b7-42f4-9644-8aeac3a7fd0f service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.378524] env[63418]: DEBUG oslo_vmware.api [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.683701} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.378868] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.379072] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.379260] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.379431] env[63418]: INFO nova.compute.manager [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Took 1.11 seconds to destroy the instance on the hypervisor. [ 823.379672] env[63418]: DEBUG oslo.service.loopingcall [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.379860] env[63418]: DEBUG nova.compute.manager [-] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 823.379952] env[63418]: DEBUG nova.network.neutron [-] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.501260] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067854} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.504381] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.508103] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9f1852-689c-41b6-986e-fda39bb521dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.532022] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 849552d4-0bb4-48ee-af7b-390183bd2189/849552d4-0bb4-48ee-af7b-390183bd2189.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.532259] env[63418]: DEBUG nova.network.neutron [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [{"id": "2024d711-87c3-497d-8373-842a53f06798", "address": "fa:16:3e:e9:da:bf", "network": {"id": "3c73f0ee-1a86-4875-8c44-5658ef2ff43f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1851341556-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c828658cd364400a9207d0c90f3ae1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2024d711-87", "ovs_interfaceid": "2024d711-87c3-497d-8373-842a53f06798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.534451] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6ab2473-8e4a-47b7-9876-6b70e13e6fd8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.550545] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b94a44-4ca9-42bc-b075-43c94184eca0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.556939] env[63418]: DEBUG oslo_concurrency.lockutils [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Releasing lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.564697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.564850] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.565089] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.565283] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.565455] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.568849] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03aec32-fe92-4f32-b2c4-3389d61fa866 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.572661] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 823.572661] env[63418]: value = "task-1245004" [ 823.572661] env[63418]: _type = "Task" [ 823.572661] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.573570] env[63418]: INFO nova.compute.manager [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Terminating instance [ 823.605718] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052fc02c-186d-44e5-aa28-a8f0ea4864b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.612177] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.617813] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc290ce-bb5e-4de2-82c0-cfa43717938e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.622738] env[63418]: DEBUG nova.network.neutron [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Successfully created port: 0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.625270] env[63418]: DEBUG nova.compute.manager [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 823.625514] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.626587] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb08098-ff85-4be9-a9ff-2d04cedd393c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.639020] env[63418]: DEBUG nova.compute.provider_tree [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 823.644157] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 823.646532] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8501bc6-d224-4760-ac68-51841c107a05 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.653072] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 823.655599] env[63418]: DEBUG oslo_vmware.api [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 823.655599] env[63418]: value = "task-1245005" [ 823.655599] env[63418]: _type = "Task" [ 823.655599] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.665541] env[63418]: DEBUG oslo_vmware.api [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1245005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.669167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.714543] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245003, 'name': Rename_Task, 'duration_secs': 0.273201} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.716322] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.716678] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5e716ba-9b15-4407-a2d6-8e0158eae36d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.723326] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 823.723326] env[63418]: value = "task-1245006" [ 823.723326] env[63418]: _type = "Task" [ 823.723326] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.733983] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.836631] env[63418]: DEBUG nova.network.neutron [-] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.872031] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc6a7ad9-5dd5-4247-a8c0-2bcc68c35cf4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.881945] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08e12ab-eb03-4025-96ce-60924bc214b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.914196] env[63418]: DEBUG nova.compute.manager [req-23f28a90-b6fa-4d88-a936-d4eddd2c632e req-ae9a8f46-d1b7-42f4-9644-8aeac3a7fd0f service nova] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Detach interface failed, port_id=9d15beed-d1a9-4234-b1ab-fd8df3c8abc6, reason: Instance 6b81dbe2-aa9e-4561-962b-2af167234b90 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 823.984792] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.061153] env[63418]: DEBUG nova.compute.manager [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 824.062026] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b8e2ea-9e6f-48f8-b196-2a2f95c2f4d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.084939] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.104860] env[63418]: DEBUG nova.compute.manager [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 824.105208] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 824.106074] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392bb4e1-8f99-4653-bbed-c67b01641432 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.114829] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 824.115089] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-beb03d65-ad7b-419e-a455-4bc3a2baf595 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.121223] env[63418]: DEBUG oslo_vmware.api [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 824.121223] env[63418]: value = "task-1245007" [ 824.121223] env[63418]: _type = "Task" [ 824.121223] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.129625] env[63418]: DEBUG oslo_vmware.api [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.164574] env[63418]: ERROR nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] [req-cff0e6b6-7d32-400d-9f22-7b459948f535] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cff0e6b6-7d32-400d-9f22-7b459948f535"}]} [ 824.171251] env[63418]: DEBUG oslo_vmware.api [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1245005, 'name': PowerOffVM_Task, 'duration_secs': 0.20518} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.171251] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 824.171251] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 824.171251] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0767656-ad20-4db3-81b3-3517bddfdf9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.187656] env[63418]: DEBUG nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 824.209229] env[63418]: DEBUG nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 824.209229] env[63418]: DEBUG nova.compute.provider_tree [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.219720] env[63418]: DEBUG nova.network.neutron [-] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.223296] env[63418]: DEBUG nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 824.238037] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245006, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.259382] env[63418]: DEBUG nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 824.268730] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 824.268987] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 824.269198] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Deleting the datastore file [datastore1] f7554ae4-c7a7-4111-a830-10f9029dc074 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.269490] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2640c158-593b-47e9-adfa-506a630b0004 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.277154] env[63418]: DEBUG oslo_vmware.api [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for the task: (returnval){ [ 824.277154] env[63418]: value = "task-1245009" [ 824.277154] env[63418]: _type = "Task" [ 824.277154] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.287035] env[63418]: DEBUG oslo_vmware.api [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1245009, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.338829] env[63418]: INFO nova.compute.manager [-] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Took 1.39 seconds to deallocate network for instance. [ 824.571541] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546c534e-6889-479e-bdd2-db882e8af956 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.583521] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af764e30-f6ac-41ba-a5ed-e8eda3b99494 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.589628] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245004, 'name': ReconfigVM_Task, 'duration_secs': 0.924216} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.590902] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 849552d4-0bb4-48ee-af7b-390183bd2189/849552d4-0bb4-48ee-af7b-390183bd2189.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.590902] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3eb4350e-4452-48a3-a1d4-61b98e091ad8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.618034] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a444a370-0baf-43a4-a3a5-e809546326d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.622608] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 824.622608] env[63418]: value = "task-1245010" [ 824.622608] env[63418]: _type = "Task" [ 824.622608] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.631258] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36fcb21c-df6a-4b95-bf59-d8311d9adc4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.638224] env[63418]: DEBUG oslo_vmware.api [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245007, 'name': PowerOffVM_Task, 'duration_secs': 0.188955} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.638765] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 824.638945] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 824.639190] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a79298b8-b40a-40bd-be80-8b0524092bbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.650955] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245010, 'name': Rename_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.651707] env[63418]: DEBUG nova.compute.provider_tree [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.664674] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 824.691641] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.691920] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.692545] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.692545] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.692545] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.692739] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.692773] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.692927] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.693125] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.693305] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.693487] env[63418]: DEBUG nova.virt.hardware [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.694413] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecaef3f-01cf-46b0-9fe2-6c734abec3d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.702961] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaabdbf6-27b1-44e0-a963-5653cc5efa56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.720441] env[63418]: INFO nova.compute.manager [-] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Took 1.34 seconds to deallocate network for instance. [ 824.737645] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245006, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.786598] env[63418]: DEBUG oslo_vmware.api [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Task: {'id': task-1245009, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140065} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.786875] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.787090] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 824.787272] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.787450] env[63418]: INFO nova.compute.manager [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Took 1.16 seconds to destroy the instance on the hypervisor. [ 824.787698] env[63418]: DEBUG oslo.service.loopingcall [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.787890] env[63418]: DEBUG nova.compute.manager [-] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 824.787986] env[63418]: DEBUG nova.network.neutron [-] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.805897] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 824.806086] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 824.806303] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleting the datastore file [datastore1] be1c134c-f36e-4cc7-b4ef-8f30793fb4df {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.806583] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e557eca4-5a01-4dbc-9b3b-0f7edcaef6ba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.813326] env[63418]: DEBUG oslo_vmware.api [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 824.813326] env[63418]: value = "task-1245012" [ 824.813326] env[63418]: _type = "Task" [ 824.813326] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.822071] env[63418]: DEBUG oslo_vmware.api [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245012, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.850716] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.080781] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4c29a1-17a7-4dc8-9f77-a247eb2a1161 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.092845] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Doing hard reboot of VM {{(pid=63418) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 825.093151] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-6f4260f3-e86b-48c0-91e8-20b53ae1033a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.099642] env[63418]: DEBUG oslo_vmware.api [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 825.099642] env[63418]: value = "task-1245013" [ 825.099642] env[63418]: _type = "Task" [ 825.099642] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.108160] env[63418]: DEBUG oslo_vmware.api [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1245013, 'name': ResetVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.136091] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245010, 'name': Rename_Task, 'duration_secs': 0.275551} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.136477] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.136793] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15e1a564-8be8-4499-a753-06a713d7ca23 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.143511] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 825.143511] env[63418]: value = "task-1245014" [ 825.143511] env[63418]: _type = "Task" [ 825.143511] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.151481] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245014, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.189286] env[63418]: DEBUG nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 825.189645] env[63418]: DEBUG nova.compute.provider_tree [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 83 to 84 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 825.189901] env[63418]: DEBUG nova.compute.provider_tree [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.226548] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.239026] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245006, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.323913] env[63418]: DEBUG oslo_vmware.api [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245012, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23125} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.324235] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.324428] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 825.324604] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.324785] env[63418]: INFO nova.compute.manager [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Took 1.22 seconds to destroy the instance on the hypervisor. [ 825.325073] env[63418]: DEBUG oslo.service.loopingcall [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.325291] env[63418]: DEBUG nova.compute.manager [-] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 825.325384] env[63418]: DEBUG nova.network.neutron [-] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.610816] env[63418]: DEBUG nova.network.neutron [-] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.612274] env[63418]: DEBUG oslo_vmware.api [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1245013, 'name': ResetVM_Task, 'duration_secs': 0.125525} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.612385] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Did hard reboot of VM {{(pid=63418) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 825.612581] env[63418]: DEBUG nova.compute.manager [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 825.613763] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77e5bbc-e9c6-4e44-a638-efc941e0ba2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.656856] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245014, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.659758] env[63418]: DEBUG nova.compute.manager [req-5f38026f-6c6f-4b82-b0e7-27f16ea5a9bf req-fb9cef48-9b66-47b7-92b1-2062c5492f4c service nova] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Received event network-vif-deleted-75354cd4-70a1-4aa8-aab4-6ae9cb843f5a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 825.660670] env[63418]: DEBUG nova.compute.manager [req-5f38026f-6c6f-4b82-b0e7-27f16ea5a9bf req-fb9cef48-9b66-47b7-92b1-2062c5492f4c service nova] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Received event network-vif-deleted-e1a45abb-8278-47dc-9780-4592d326c13d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 825.697315] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.053s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.698438] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.445s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.698781] env[63418]: DEBUG nova.objects.instance [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lazy-loading 'resources' on Instance uuid e3dd6d5e-2b80-4724-9826-6ebfb8181c3b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.741208] env[63418]: DEBUG oslo_vmware.api [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245006, 'name': PowerOnVM_Task, 'duration_secs': 1.653805} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.743322] env[63418]: INFO nova.scheduler.client.report [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Deleted allocations for instance 27290558-6c58-414d-bd53-c4a686890721 [ 825.746104] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.746104] env[63418]: INFO nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Took 11.15 seconds to spawn the instance on the hypervisor. [ 825.746104] env[63418]: DEBUG nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 825.750221] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290a172d-e75e-4514-b445-d3f315f05e4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.767254] env[63418]: DEBUG nova.compute.manager [req-32917df8-5526-4b95-be34-da4c24b65bd2 req-26ed55be-51ab-4ebe-9593-c302736a2ed4 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Received event network-vif-plugged-0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 825.767254] env[63418]: DEBUG oslo_concurrency.lockutils [req-32917df8-5526-4b95-be34-da4c24b65bd2 req-26ed55be-51ab-4ebe-9593-c302736a2ed4 service nova] Acquiring lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.767254] env[63418]: DEBUG oslo_concurrency.lockutils [req-32917df8-5526-4b95-be34-da4c24b65bd2 req-26ed55be-51ab-4ebe-9593-c302736a2ed4 service nova] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.767254] env[63418]: DEBUG oslo_concurrency.lockutils [req-32917df8-5526-4b95-be34-da4c24b65bd2 req-26ed55be-51ab-4ebe-9593-c302736a2ed4 service nova] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.767254] env[63418]: DEBUG nova.compute.manager [req-32917df8-5526-4b95-be34-da4c24b65bd2 req-26ed55be-51ab-4ebe-9593-c302736a2ed4 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] No waiting events found dispatching network-vif-plugged-0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 825.767626] env[63418]: WARNING nova.compute.manager [req-32917df8-5526-4b95-be34-da4c24b65bd2 req-26ed55be-51ab-4ebe-9593-c302736a2ed4 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Received unexpected event network-vif-plugged-0591996d-662a-4dac-80d1-c81781f6e17d for instance with vm_state building and task_state spawning. [ 825.858610] env[63418]: DEBUG nova.network.neutron [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Successfully updated port: 0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.069982] env[63418]: DEBUG nova.network.neutron [-] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.114154] env[63418]: INFO nova.compute.manager [-] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Took 1.32 seconds to deallocate network for instance. [ 826.131874] env[63418]: DEBUG oslo_concurrency.lockutils [None req-faa1f56a-745d-458b-8c4d-83c90b9be7ee tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.088s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.155428] env[63418]: DEBUG oslo_vmware.api [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245014, 'name': PowerOnVM_Task, 'duration_secs': 1.007778} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.155699] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.155904] env[63418]: INFO nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Took 9.05 seconds to spawn the instance on the hypervisor. [ 826.156100] env[63418]: DEBUG nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 826.156890] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdcaa30-e00a-43f7-9a5b-b2c2d69e40cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.255673] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7bfd2e3d-6b9c-4f52-8eee-e136273aece7 tempest-ServerAddressesNegativeTestJSON-1238411037 tempest-ServerAddressesNegativeTestJSON-1238411037-project-member] Lock "27290558-6c58-414d-bd53-c4a686890721" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.328s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.279610] env[63418]: INFO nova.compute.manager [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Took 33.26 seconds to build instance. [ 826.363309] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.363466] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.363586] env[63418]: DEBUG nova.network.neutron [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.534562] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a65804-7be1-45ea-bd0c-c1ed0d45fca7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.543168] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae1b809-94ed-4989-94e1-79d072c1cf58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.576017] env[63418]: INFO nova.compute.manager [-] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Took 1.25 seconds to deallocate network for instance. [ 826.577633] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5068e2a-d537-4765-be3d-ccb2a8878189 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.591815] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e498fb1-e256-4be8-99a9-a3bb02c6671a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.602673] env[63418]: DEBUG nova.compute.provider_tree [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.619093] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.674842] env[63418]: INFO nova.compute.manager [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Took 33.24 seconds to build instance. [ 826.782404] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3928d156-f547-4d9c-990b-76f7a58dee48 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.281s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.924106] env[63418]: DEBUG nova.network.neutron [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.085979] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.105887] env[63418]: DEBUG nova.scheduler.client.report [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 827.117440] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.117672] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.176664] env[63418]: DEBUG oslo_concurrency.lockutils [None req-56566819-92c2-409e-ba52-095af8b6965c tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "849552d4-0bb4-48ee-af7b-390183bd2189" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.564s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.218021] env[63418]: DEBUG nova.network.neutron [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updating instance_info_cache with network_info: [{"id": "0591996d-662a-4dac-80d1-c81781f6e17d", "address": "fa:16:3e:d5:0b:db", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0591996d-66", "ovs_interfaceid": "0591996d-662a-4dac-80d1-c81781f6e17d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.615672] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.619472] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.367s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.622261] env[63418]: INFO nova.compute.claims [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.625798] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 827.658925] env[63418]: INFO nova.scheduler.client.report [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Deleted allocations for instance e3dd6d5e-2b80-4724-9826-6ebfb8181c3b [ 827.723719] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.724124] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Instance network_info: |[{"id": "0591996d-662a-4dac-80d1-c81781f6e17d", "address": "fa:16:3e:d5:0b:db", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0591996d-66", "ovs_interfaceid": "0591996d-662a-4dac-80d1-c81781f6e17d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 827.725039] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:0b:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0591996d-662a-4dac-80d1-c81781f6e17d', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.734334] env[63418]: DEBUG oslo.service.loopingcall [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.735020] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.736070] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6c8ef0e-4127-494f-99bd-1458ea8751e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.761387] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.761387] env[63418]: value = "task-1245015" [ 827.761387] env[63418]: _type = "Task" [ 827.761387] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.769043] env[63418]: DEBUG nova.compute.manager [req-3feee3d9-a60d-4e77-a524-bf45215d2923 req-dd546863-6f18-4f2d-ae62-71201aa9e9b1 service nova] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Received event network-vif-deleted-48c02e8f-dada-4b14-bee4-8bee405a7471 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 827.773693] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245015, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.860555] env[63418]: DEBUG nova.compute.manager [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Received event network-changed-0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 827.860985] env[63418]: DEBUG nova.compute.manager [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Refreshing instance network info cache due to event network-changed-0591996d-662a-4dac-80d1-c81781f6e17d. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 827.861342] env[63418]: DEBUG oslo_concurrency.lockutils [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] Acquiring lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.861597] env[63418]: DEBUG oslo_concurrency.lockutils [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] Acquired lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.861861] env[63418]: DEBUG nova.network.neutron [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Refreshing network info cache for port 0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.152607] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.168277] env[63418]: DEBUG oslo_concurrency.lockutils [None req-544a1033-c734-4c56-a61c-33f9a91ab90f tempest-ServerRescueTestJSONUnderV235-1742639264 tempest-ServerRescueTestJSONUnderV235-1742639264-project-member] Lock "e3dd6d5e-2b80-4724-9826-6ebfb8181c3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.639s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.258468] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.258776] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.259718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "fa4c0315-1ef5-4491-ab18-d49563b778fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.259718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.259718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.265659] env[63418]: INFO nova.compute.manager [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Terminating instance [ 828.274417] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245015, 'name': CreateVM_Task, 'duration_secs': 0.379717} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.274736] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.275626] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.277167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.277167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.277310] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fd1ed7e-0456-4432-981e-f700b9e6a0b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.285018] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 828.285018] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5225b674-0036-4d3a-cc86-4d59b99fc67a" [ 828.285018] env[63418]: _type = "Task" [ 828.285018] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.292324] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5225b674-0036-4d3a-cc86-4d59b99fc67a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.765527] env[63418]: DEBUG nova.network.neutron [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updated VIF entry in instance network info cache for port 0591996d-662a-4dac-80d1-c81781f6e17d. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 828.768874] env[63418]: DEBUG nova.network.neutron [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updating instance_info_cache with network_info: [{"id": "0591996d-662a-4dac-80d1-c81781f6e17d", "address": "fa:16:3e:d5:0b:db", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0591996d-66", "ovs_interfaceid": "0591996d-662a-4dac-80d1-c81781f6e17d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.770712] env[63418]: DEBUG nova.compute.manager [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 828.770952] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.776632] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d49d1b-0b04-4e10-a480-738a9d609d5d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.797851] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 828.798681] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0a00f9b-0fbe-4263-a8bb-e3a59005d8c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.810063] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5225b674-0036-4d3a-cc86-4d59b99fc67a, 'name': SearchDatastore_Task, 'duration_secs': 0.010721} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.810845] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.811530] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.811530] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.811714] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.811764] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.812291] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d10ddd66-abf6-492f-8b8a-1203ec80b789 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.816511] env[63418]: DEBUG oslo_vmware.api [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 828.816511] env[63418]: value = "task-1245016" [ 828.816511] env[63418]: _type = "Task" [ 828.816511] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.825969] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.826822] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.828511] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c54a865c-fb38-400d-80b8-d45470caa884 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.834200] env[63418]: DEBUG oslo_vmware.api [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1245016, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.837414] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 828.837414] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5213200a-b48e-8969-5f7f-d54264048f09" [ 828.837414] env[63418]: _type = "Task" [ 828.837414] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.849635] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5213200a-b48e-8969-5f7f-d54264048f09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.006601] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439bee12-c61d-439e-ae02-d17c0cc35430 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.016236] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65df3624-7d30-41da-9be3-6538388b7e67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.053396] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7deaff34-c650-48ae-9390-828de46a7b34 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.061826] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a615509d-b671-49b2-9351-4cd6e193cdca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.076772] env[63418]: DEBUG nova.compute.provider_tree [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.273569] env[63418]: DEBUG oslo_concurrency.lockutils [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] Releasing lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.273882] env[63418]: DEBUG nova.compute.manager [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Received event network-changed-2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 829.274086] env[63418]: DEBUG nova.compute.manager [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Refreshing instance network info cache due to event network-changed-2024d711-87c3-497d-8373-842a53f06798. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 829.274316] env[63418]: DEBUG oslo_concurrency.lockutils [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] Acquiring lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.275406] env[63418]: DEBUG oslo_concurrency.lockutils [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] Acquired lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.275406] env[63418]: DEBUG nova.network.neutron [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Refreshing network info cache for port 2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.285900] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "849552d4-0bb4-48ee-af7b-390183bd2189" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.286154] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "849552d4-0bb4-48ee-af7b-390183bd2189" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.286372] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "849552d4-0bb4-48ee-af7b-390183bd2189-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.286555] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "849552d4-0bb4-48ee-af7b-390183bd2189-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.286721] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "849552d4-0bb4-48ee-af7b-390183bd2189-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.288925] env[63418]: INFO nova.compute.manager [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Terminating instance [ 829.330741] env[63418]: DEBUG oslo_vmware.api [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1245016, 'name': PowerOffVM_Task, 'duration_secs': 0.433436} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.331197] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 829.331449] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.331960] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63b4d081-595d-403e-9530-b4588538031c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.347530] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5213200a-b48e-8969-5f7f-d54264048f09, 'name': SearchDatastore_Task, 'duration_secs': 0.008354} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.348330] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc3a840f-1ba1-4df6-b9e5-f3d964860a9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.354092] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 829.354092] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526bdabb-a631-487a-aadc-cb8f63433dc1" [ 829.354092] env[63418]: _type = "Task" [ 829.354092] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.362546] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526bdabb-a631-487a-aadc-cb8f63433dc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.427088] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.427573] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.427828] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Deleting the datastore file [datastore1] fa4c0315-1ef5-4491-ab18-d49563b778fb {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.428744] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6d4dd15-2a87-423e-bd40-be71b5b6f6a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.436653] env[63418]: DEBUG oslo_vmware.api [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for the task: (returnval){ [ 829.436653] env[63418]: value = "task-1245018" [ 829.436653] env[63418]: _type = "Task" [ 829.436653] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.446315] env[63418]: DEBUG oslo_vmware.api [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1245018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.580518] env[63418]: DEBUG nova.scheduler.client.report [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 829.792197] env[63418]: DEBUG nova.compute.manager [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 829.792449] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.793411] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-143e6cb2-93c8-4dda-ac8d-e014973651a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.802442] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.802729] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70ce1981-779a-4e6d-969f-1e188c9e0a4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.812317] env[63418]: DEBUG oslo_vmware.api [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 829.812317] env[63418]: value = "task-1245019" [ 829.812317] env[63418]: _type = "Task" [ 829.812317] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.822283] env[63418]: DEBUG oslo_vmware.api [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.865385] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526bdabb-a631-487a-aadc-cb8f63433dc1, 'name': SearchDatastore_Task, 'duration_secs': 0.008841} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.865932] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.866326] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 61af37a5-8c18-47a7-817a-eb3b332b3725/61af37a5-8c18-47a7-817a-eb3b332b3725.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.868914] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0bf0196-5305-4e02-a719-08fe595eb148 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.876105] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 829.876105] env[63418]: value = "task-1245020" [ 829.876105] env[63418]: _type = "Task" [ 829.876105] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.885385] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.949282] env[63418]: DEBUG oslo_vmware.api [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Task: {'id': task-1245018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126483} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.949758] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.950247] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.950247] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.950325] env[63418]: INFO nova.compute.manager [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Took 1.18 seconds to destroy the instance on the hypervisor. [ 829.950570] env[63418]: DEBUG oslo.service.loopingcall [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.950812] env[63418]: DEBUG nova.compute.manager [-] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 829.950921] env[63418]: DEBUG nova.network.neutron [-] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 830.077853] env[63418]: DEBUG nova.network.neutron [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updated VIF entry in instance network info cache for port 2024d711-87c3-497d-8373-842a53f06798. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 830.078336] env[63418]: DEBUG nova.network.neutron [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [{"id": "2024d711-87c3-497d-8373-842a53f06798", "address": "fa:16:3e:e9:da:bf", "network": {"id": "3c73f0ee-1a86-4875-8c44-5658ef2ff43f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1851341556-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c828658cd364400a9207d0c90f3ae1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2024d711-87", "ovs_interfaceid": "2024d711-87c3-497d-8373-842a53f06798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.088031] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.088592] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 830.091271] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.238s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.091553] env[63418]: DEBUG nova.objects.instance [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lazy-loading 'resources' on Instance uuid 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.322274] env[63418]: DEBUG oslo_vmware.api [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245019, 'name': PowerOffVM_Task, 'duration_secs': 0.207042} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.322622] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.322723] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.323829] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f233f67-cf0c-40df-bb58-908a9abdb2db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.385365] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.385596] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.385783] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Deleting the datastore file [datastore2] 849552d4-0bb4-48ee-af7b-390183bd2189 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.388155] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d7edfe0-d474-468d-9f92-a7aee0b002dc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.401484] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245020, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.407358] env[63418]: DEBUG oslo_vmware.api [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for the task: (returnval){ [ 830.407358] env[63418]: value = "task-1245022" [ 830.407358] env[63418]: _type = "Task" [ 830.407358] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.419328] env[63418]: DEBUG oslo_vmware.api [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.479611] env[63418]: DEBUG nova.compute.manager [req-33604371-27ae-4a50-be69-ad188bc13df8 req-4f2f1d16-a967-4678-a5a7-a3312b690f5f service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Received event network-vif-deleted-2024d711-87c3-497d-8373-842a53f06798 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 830.480248] env[63418]: INFO nova.compute.manager [req-33604371-27ae-4a50-be69-ad188bc13df8 req-4f2f1d16-a967-4678-a5a7-a3312b690f5f service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Neutron deleted interface 2024d711-87c3-497d-8373-842a53f06798; detaching it from the instance and deleting it from the info cache [ 830.480572] env[63418]: DEBUG nova.network.neutron [req-33604371-27ae-4a50-be69-ad188bc13df8 req-4f2f1d16-a967-4678-a5a7-a3312b690f5f service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.581082] env[63418]: DEBUG oslo_concurrency.lockutils [req-3f54a508-87a2-45fd-a66f-1e357840a6b3 req-2eba5c39-f0a1-4585-bb02-7b22a799221a service nova] Releasing lock "refresh_cache-fa4c0315-1ef5-4491-ab18-d49563b778fb" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.597021] env[63418]: DEBUG nova.compute.utils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.597021] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 830.597021] env[63418]: DEBUG nova.network.neutron [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.654718] env[63418]: DEBUG nova.policy [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ef8f62bd75a4d109a2e1881408754cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11fffb9badce4abeadce3ab70dff7d58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.794803] env[63418]: DEBUG nova.network.neutron [-] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.890954] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245020, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565859} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.891327] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 61af37a5-8c18-47a7-817a-eb3b332b3725/61af37a5-8c18-47a7-817a-eb3b332b3725.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.891486] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.895877] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0572e1f5-a3e7-49ab-bbde-fa4c50999e90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.902589] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 830.902589] env[63418]: value = "task-1245023" [ 830.902589] env[63418]: _type = "Task" [ 830.902589] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.909411] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee697d89-ef86-4838-9595-90461229a15e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.920777] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245023, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.927834] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddbd8c3-9a3f-4b42-89fd-df52ae88bb12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.933834] env[63418]: DEBUG oslo_vmware.api [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Task: {'id': task-1245022, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146729} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.933834] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.933834] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 830.933834] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 830.933834] env[63418]: INFO nova.compute.manager [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Took 1.14 seconds to destroy the instance on the hypervisor. [ 830.934059] env[63418]: DEBUG oslo.service.loopingcall [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.934059] env[63418]: DEBUG nova.compute.manager [-] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 830.934059] env[63418]: DEBUG nova.network.neutron [-] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 830.964145] env[63418]: INFO nova.compute.manager [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Rebuilding instance [ 830.969469] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ee2546-4a10-4dc6-a298-3b96655ffcef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.971254] env[63418]: DEBUG nova.network.neutron [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Successfully created port: bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.978297] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3e5121-859c-4a07-ae91-145ffc727333 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.991644] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c145b72e-8766-439c-86df-26aa6d7cbf83 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.993853] env[63418]: DEBUG nova.compute.provider_tree [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.003780] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ff64be-e6e3-442d-9581-8210e369c61c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.040866] env[63418]: DEBUG nova.compute.manager [req-33604371-27ae-4a50-be69-ad188bc13df8 req-4f2f1d16-a967-4678-a5a7-a3312b690f5f service nova] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Detach interface failed, port_id=2024d711-87c3-497d-8373-842a53f06798, reason: Instance fa4c0315-1ef5-4491-ab18-d49563b778fb could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 831.044296] env[63418]: DEBUG nova.compute.manager [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 831.045101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3a7d98-30c5-411e-910e-98aab7d22a8e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.102089] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 831.298107] env[63418]: INFO nova.compute.manager [-] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Took 1.35 seconds to deallocate network for instance. [ 831.412110] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245023, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073425} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.412596] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.413142] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcc3828-0574-4e0b-b1c3-950b03da4f53 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.439182] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 61af37a5-8c18-47a7-817a-eb3b332b3725/61af37a5-8c18-47a7-817a-eb3b332b3725.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.439707] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9c9e9bb-a315-4c48-bcbf-3c4970a079e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.464937] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 831.464937] env[63418]: value = "task-1245024" [ 831.464937] env[63418]: _type = "Task" [ 831.464937] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.473614] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245024, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.500273] env[63418]: DEBUG nova.scheduler.client.report [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 831.604696] env[63418]: DEBUG nova.compute.manager [req-0aa0d92e-2ad0-4e53-be4c-ff84ef7ad200 req-ff3b7e71-d8b2-42a0-a9f7-3bece42f2456 service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Received event network-vif-deleted-43c32def-abd1-4085-95c5-a30bfcbcf4e0 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 831.604943] env[63418]: INFO nova.compute.manager [req-0aa0d92e-2ad0-4e53-be4c-ff84ef7ad200 req-ff3b7e71-d8b2-42a0-a9f7-3bece42f2456 service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Neutron deleted interface 43c32def-abd1-4085-95c5-a30bfcbcf4e0; detaching it from the instance and deleting it from the info cache [ 831.605626] env[63418]: DEBUG nova.network.neutron [req-0aa0d92e-2ad0-4e53-be4c-ff84ef7ad200 req-ff3b7e71-d8b2-42a0-a9f7-3bece42f2456 service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.805133] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.974744] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245024, 'name': ReconfigVM_Task, 'duration_secs': 0.350112} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.975038] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 61af37a5-8c18-47a7-817a-eb3b332b3725/61af37a5-8c18-47a7-817a-eb3b332b3725.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.975689] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9aa55c2c-9d61-435c-b4ee-a8cce0a0faa6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.982342] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 831.982342] env[63418]: value = "task-1245025" [ 831.982342] env[63418]: _type = "Task" [ 831.982342] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.990227] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245025, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.004201] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.006253] env[63418]: DEBUG nova.network.neutron [-] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.007616] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.594s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.007843] env[63418]: DEBUG nova.objects.instance [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lazy-loading 'resources' on Instance uuid 0c151b0c-c383-4421-b506-b7afa95e2072 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.037037] env[63418]: INFO nova.scheduler.client.report [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocations for instance 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c [ 832.058831] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.059329] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c5902f1-7089-4aed-a69f-2b7020db91ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.066451] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 832.066451] env[63418]: value = "task-1245026" [ 832.066451] env[63418]: _type = "Task" [ 832.066451] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.074696] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245026, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.109157] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08cf79a2-349b-4b0c-91cd-0adcc75f7a78 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.112370] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 832.122215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bea4471-e56b-41fa-824a-7cda2ae6148f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.141047] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.141236] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.141335] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.141456] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.141606] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.141759] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.141985] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.142187] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.142365] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.142534] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.142708] env[63418]: DEBUG nova.virt.hardware [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.143573] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1138c5-d7fa-4647-ad6b-bdcbc58d5d0b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.155657] env[63418]: DEBUG nova.compute.manager [req-0aa0d92e-2ad0-4e53-be4c-ff84ef7ad200 req-ff3b7e71-d8b2-42a0-a9f7-3bece42f2456 service nova] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Detach interface failed, port_id=43c32def-abd1-4085-95c5-a30bfcbcf4e0, reason: Instance 849552d4-0bb4-48ee-af7b-390183bd2189 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 832.161432] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04077f01-7691-4780-b085-546d4f83f18c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.479574] env[63418]: DEBUG nova.network.neutron [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Successfully updated port: bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.493012] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245025, 'name': Rename_Task, 'duration_secs': 0.140745} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.494079] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.494313] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20af7452-d415-45c9-b28a-ad6c2d9d6b9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.501464] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 832.501464] env[63418]: value = "task-1245027" [ 832.501464] env[63418]: _type = "Task" [ 832.501464] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.510504] env[63418]: INFO nova.compute.manager [-] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Took 1.58 seconds to deallocate network for instance. [ 832.513153] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245027, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.517456] env[63418]: DEBUG nova.compute.manager [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Received event network-vif-plugged-bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 832.517708] env[63418]: DEBUG oslo_concurrency.lockutils [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] Acquiring lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.517914] env[63418]: DEBUG oslo_concurrency.lockutils [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.518098] env[63418]: DEBUG oslo_concurrency.lockutils [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.518279] env[63418]: DEBUG nova.compute.manager [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] No waiting events found dispatching network-vif-plugged-bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 832.518448] env[63418]: WARNING nova.compute.manager [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Received unexpected event network-vif-plugged-bf027be2-ee92-4c69-ad07-fe9393549d89 for instance with vm_state building and task_state spawning. [ 832.518611] env[63418]: DEBUG nova.compute.manager [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Received event network-changed-bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 832.518767] env[63418]: DEBUG nova.compute.manager [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Refreshing instance network info cache due to event network-changed-bf027be2-ee92-4c69-ad07-fe9393549d89. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 832.518955] env[63418]: DEBUG oslo_concurrency.lockutils [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] Acquiring lock "refresh_cache-e2dcb8b8-e778-4202-8808-6a8535e2f1b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.519109] env[63418]: DEBUG oslo_concurrency.lockutils [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] Acquired lock "refresh_cache-e2dcb8b8-e778-4202-8808-6a8535e2f1b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.519266] env[63418]: DEBUG nova.network.neutron [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Refreshing network info cache for port bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.546292] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40780537-7b16-4cf1-85c8-080d4e563141 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1f898bd4-0b80-48c1-9e7c-8948f4b19c0c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.427s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.577204] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245026, 'name': PowerOffVM_Task, 'duration_secs': 0.21902} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.580794] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.581785] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.582409] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a0512fe-2290-403c-bffa-4950d036040a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.589398] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 832.589398] env[63418]: value = "task-1245028" [ 832.589398] env[63418]: _type = "Task" [ 832.589398] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.601806] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 832.602127] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 832.602401] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268382', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'name': 'volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'serial': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 832.603648] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa9cf85-2fe4-485d-a169-fa0b79bb325a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.628635] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9255982e-d8ab-4b0a-b58d-d25a72b4bccf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.636114] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16128bab-d1a7-4d1b-94b0-8c611d229e6a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.659365] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b069d7-8cf0-4213-9004-a9fe5d9b8772 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.675736] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] The volume has not been displaced from its original location: [datastore1] volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce/volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 832.681145] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Reconfiguring VM instance instance-00000039 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 832.684081] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58c899a4-b916-4c83-8c8b-e3aadb3d2791 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.702933] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 832.702933] env[63418]: value = "task-1245029" [ 832.702933] env[63418]: _type = "Task" [ 832.702933] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.711371] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245029, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.853583] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9efef1-4543-4a9e-8f03-e2673521f202 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.861253] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404411b3-26b0-4a00-b8af-ee8cbd598b61 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.891494] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a10fc95-2469-4666-92c6-00147a114a7a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.899228] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b3a582-7862-4544-9bd1-c6b63566299c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.912683] env[63418]: DEBUG nova.compute.provider_tree [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.982368] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-e2dcb8b8-e778-4202-8808-6a8535e2f1b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.013149] env[63418]: DEBUG oslo_vmware.api [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245027, 'name': PowerOnVM_Task, 'duration_secs': 0.508146} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.013455] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.013684] env[63418]: INFO nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Took 8.35 seconds to spawn the instance on the hypervisor. [ 833.013885] env[63418]: DEBUG nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 833.014701] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c3be2f-6e9b-4fb2-8209-c5830297806e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.024024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.056332] env[63418]: DEBUG nova.network.neutron [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.173446] env[63418]: DEBUG nova.network.neutron [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.213163] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245029, 'name': ReconfigVM_Task, 'duration_secs': 0.18145} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.213501] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Reconfigured VM instance instance-00000039 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 833.218815] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d01d670f-5cb2-44c8-8eec-8d4e5cf1498e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.237641] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 833.237641] env[63418]: value = "task-1245030" [ 833.237641] env[63418]: _type = "Task" [ 833.237641] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.245765] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245030, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.415925] env[63418]: DEBUG nova.scheduler.client.report [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 833.534893] env[63418]: INFO nova.compute.manager [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Took 29.48 seconds to build instance. [ 833.676410] env[63418]: DEBUG oslo_concurrency.lockutils [req-ac9f069c-f2b3-4dd7-99b6-7cfd52274ad6 req-01f65f70-9281-4d42-b74c-ac0ff45945bd service nova] Releasing lock "refresh_cache-e2dcb8b8-e778-4202-8808-6a8535e2f1b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.676818] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-e2dcb8b8-e778-4202-8808-6a8535e2f1b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.676991] env[63418]: DEBUG nova.network.neutron [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.747922] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245030, 'name': ReconfigVM_Task, 'duration_secs': 0.137205} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.748255] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268382', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'name': 'volume-bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5', 'attached_at': '', 'detached_at': '', 'volume_id': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce', 'serial': 'bef2d2e5-3fa7-48bf-8c27-f4f988f36dce'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 833.748544] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.749340] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9cd102-73fe-4069-9d5c-6f82b134a6fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.756416] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.756704] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bc3b767-a0f0-45a8-a311-aa41b9ecdb25 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.828318] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.828518] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.828705] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Deleting the datastore file [datastore1] 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.828980] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c251f813-ab5f-4d4d-b303-a3c5dd28b80f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.836193] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for the task: (returnval){ [ 833.836193] env[63418]: value = "task-1245032" [ 833.836193] env[63418]: _type = "Task" [ 833.836193] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.844159] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245032, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.921474] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.923895] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.506s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.924233] env[63418]: DEBUG nova.objects.instance [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lazy-loading 'resources' on Instance uuid f632d71d-498f-4914-9895-8f37187a295f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.943027] env[63418]: INFO nova.scheduler.client.report [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleted allocations for instance 0c151b0c-c383-4421-b506-b7afa95e2072 [ 834.037432] env[63418]: DEBUG oslo_concurrency.lockutils [None req-704c81cc-5a0b-4211-b511-f0a95e4d75f8 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.992s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.345636] env[63418]: DEBUG oslo_vmware.api [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Task: {'id': task-1245032, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.083087} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.345904] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.346111] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.346295] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.379946] env[63418]: DEBUG nova.network.neutron [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.416410] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 834.416779] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53f2c03b-edec-4a8b-b318-6ed249413bd9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.428552] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9fe636-12a2-4b56-9994-6d726d3e479a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.459854] env[63418]: ERROR nova.compute.manager [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Failed to detach volume bef2d2e5-3fa7-48bf-8c27-f4f988f36dce from /dev/sda: nova.exception.InstanceNotFound: Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 could not be found. [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Traceback (most recent call last): [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self.driver.rebuild(**kwargs) [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] raise NotImplementedError() [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] NotImplementedError [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] During handling of the above exception, another exception occurred: [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Traceback (most recent call last): [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 834.459854] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self.driver.detach_volume(context, old_connection_info, [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] return self._volumeops.detach_volume(connection_info, instance) [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._detach_volume_vmdk(connection_info, instance) [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] stable_ref.fetch_moref(session) [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] raise exception.InstanceNotFound(instance_id=self._uuid) [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] nova.exception.InstanceNotFound: Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 could not be found. [ 834.460227] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.462298] env[63418]: DEBUG oslo_concurrency.lockutils [None req-95233ecf-52ab-4275-b80e-1b789e8f8ab5 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "0c151b0c-c383-4421-b506-b7afa95e2072" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.860s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.555126] env[63418]: DEBUG nova.network.neutron [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Updating instance_info_cache with network_info: [{"id": "bf027be2-ee92-4c69-ad07-fe9393549d89", "address": "fa:16:3e:5c:4e:1d", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf027be2-ee", "ovs_interfaceid": "bf027be2-ee92-4c69-ad07-fe9393549d89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.599880] env[63418]: DEBUG nova.compute.utils [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Build of instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 aborted: Failed to rebuild volume backed instance. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.603554] env[63418]: ERROR nova.compute.manager [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 aborted: Failed to rebuild volume backed instance. [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Traceback (most recent call last): [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self.driver.rebuild(**kwargs) [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] raise NotImplementedError() [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] NotImplementedError [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] During handling of the above exception, another exception occurred: [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Traceback (most recent call last): [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3642, in _rebuild_volume_backed_instance [ 834.603554] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._detach_root_volume(context, instance, root_bdm) [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _detach_root_volume [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] with excutils.save_and_reraise_exception(): [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self.force_reraise() [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] raise self.value [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self.driver.detach_volume(context, old_connection_info, [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] return self._volumeops.detach_volume(connection_info, instance) [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 834.603998] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._detach_volume_vmdk(connection_info, instance) [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] stable_ref.fetch_moref(session) [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] raise exception.InstanceNotFound(instance_id=self._uuid) [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] nova.exception.InstanceNotFound: Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 could not be found. [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] During handling of the above exception, another exception occurred: [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Traceback (most recent call last): [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 11215, in _error_out_instance_on_exception [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] yield [ 834.604365] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3910, in rebuild_instance [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._do_rebuild_instance_with_claim( [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3996, in _do_rebuild_instance_with_claim [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._do_rebuild_instance( [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 4188, in _do_rebuild_instance [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._rebuild_default_impl(**kwargs) [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3765, in _rebuild_default_impl [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] self._rebuild_volume_backed_instance( [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] File "/opt/stack/nova/nova/compute/manager.py", line 3657, in _rebuild_volume_backed_instance [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] raise exception.BuildAbortException( [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] nova.exception.BuildAbortException: Build of instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 aborted: Failed to rebuild volume backed instance. [ 834.604813] env[63418]: ERROR nova.compute.manager [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] [ 834.652888] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 834.655604] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 834.702372] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578ee59c-6611-48f6-a400-2a4a482e4afc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.710905] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b0edf1-7763-4cff-8a3f-5280be61ec9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.739996] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f5834b-09c8-4d0a-a7a0-f7fac56e3dcb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.747185] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e855f8a-601d-4c50-8141-4a9ec4d1445e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.762059] env[63418]: DEBUG nova.compute.provider_tree [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.057841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-e2dcb8b8-e778-4202-8808-6a8535e2f1b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.058282] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance network_info: |[{"id": "bf027be2-ee92-4c69-ad07-fe9393549d89", "address": "fa:16:3e:5c:4e:1d", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf027be2-ee", "ovs_interfaceid": "bf027be2-ee92-4c69-ad07-fe9393549d89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 835.058776] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:4e:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf027be2-ee92-4c69-ad07-fe9393549d89', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.066631] env[63418]: DEBUG oslo.service.loopingcall [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.066889] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.067229] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-643f1f61-cfb9-469f-ad41-1df51a8bc421 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.086685] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.086685] env[63418]: value = "task-1245033" [ 835.086685] env[63418]: _type = "Task" [ 835.086685] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.094478] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245033, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.142500] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "61af37a5-8c18-47a7-817a-eb3b332b3725" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.142756] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.142938] env[63418]: INFO nova.compute.manager [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Shelving [ 835.162688] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 835.162840] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 835.162959] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Rebuilding the list of instances to heal {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10282}} [ 835.266268] env[63418]: DEBUG nova.scheduler.client.report [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 835.597172] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245033, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.668927] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 835.685395] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.685532] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquired lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.685670] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Forcefully refreshing network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 835.685815] env[63418]: DEBUG nova.objects.instance [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lazy-loading 'info_cache' on Instance uuid 6b81dbe2-aa9e-4561-962b-2af167234b90 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.773419] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.775903] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.550s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.776384] env[63418]: DEBUG nova.objects.instance [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'resources' on Instance uuid d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.795028] env[63418]: INFO nova.scheduler.client.report [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Deleted allocations for instance f632d71d-498f-4914-9895-8f37187a295f [ 836.097734] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245033, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.154070] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.154070] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2b169f0-412e-44a7-8c29-33e2ee7e8d03 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.161042] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 836.161042] env[63418]: value = "task-1245034" [ 836.161042] env[63418]: _type = "Task" [ 836.161042] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.167920] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.278982] env[63418]: DEBUG nova.objects.instance [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'numa_topology' on Instance uuid d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.301691] env[63418]: DEBUG oslo_concurrency.lockutils [None req-585babb2-ec84-411d-a906-4ffb69c45005 tempest-ListServersNegativeTestJSON-451409934 tempest-ListServersNegativeTestJSON-451409934-project-member] Lock "f632d71d-498f-4914-9895-8f37187a295f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.605s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.603224] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245033, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.620697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.670205] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.710592] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.781522] env[63418]: DEBUG nova.objects.base [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 837.008814] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b094d365-35f0-404b-9860-7a2e5ea56d4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.016794] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79a23d1-f507-471c-aeef-04ecb6c42000 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.054522] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b765c97f-eab4-43a7-a28a-0a14f72cee7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.063513] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a33018-b77e-4c86-8506-cf1bc9c4cc9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.079667] env[63418]: DEBUG nova.compute.provider_tree [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.097743] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245033, 'name': CreateVM_Task, 'duration_secs': 1.537346} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.097907] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.098585] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.098755] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.099113] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.099377] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2425e41-0c73-41c2-8c4d-7adcb2495de4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.103997] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 837.103997] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5262b1ca-eff6-ebe2-17b5-4ff2f05f67cd" [ 837.103997] env[63418]: _type = "Task" [ 837.103997] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.112604] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5262b1ca-eff6-ebe2-17b5-4ff2f05f67cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.170555] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245034, 'name': PowerOffVM_Task, 'duration_secs': 0.596054} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.170888] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 837.171716] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f973d89c-6b26-48e4-84d6-2a4aad399956 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.190606] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87e9dd9-c760-42dc-bc32-2044512a477a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.290443] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.290614] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.290900] env[63418]: DEBUG nova.objects.instance [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'flavor' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.293328] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.558006] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.558301] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.558523] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.558747] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.558881] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.561928] env[63418]: INFO nova.compute.manager [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Terminating instance [ 837.582402] env[63418]: DEBUG nova.scheduler.client.report [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 837.614244] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5262b1ca-eff6-ebe2-17b5-4ff2f05f67cd, 'name': SearchDatastore_Task, 'duration_secs': 0.009869} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.614546] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.614780] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.615540] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.615540] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.615540] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.615684] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2bad48d-2262-47b3-b23b-708e1afeacb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.625272] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.625782] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.626315] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cad8fe20-8697-4283-b151-573e856b6664 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.634986] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 837.634986] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529bf5b3-ad7a-7dcf-c9dd-be23bcfd933d" [ 837.634986] env[63418]: _type = "Task" [ 837.634986] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.647508] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529bf5b3-ad7a-7dcf-c9dd-be23bcfd933d, 'name': SearchDatastore_Task, 'duration_secs': 0.009253} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.648318] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74c86bc2-f82a-4723-b3a6-73b81b4b87a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.654522] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 837.654522] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529d4232-ad53-139c-1c35-049225683b60" [ 837.654522] env[63418]: _type = "Task" [ 837.654522] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.665104] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529d4232-ad53-139c-1c35-049225683b60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.701891] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 837.702253] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-64881f72-6b47-4e1b-81c4-cbe54717d7c2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.711678] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 837.711678] env[63418]: value = "task-1245035" [ 837.711678] env[63418]: _type = "Task" [ 837.711678] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.720751] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245035, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.799244] env[63418]: DEBUG nova.objects.instance [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'pci_requests' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.803437] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Releasing lock "refresh_cache-6b81dbe2-aa9e-4561-962b-2af167234b90" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.803437] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Updated the network info_cache for instance {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10349}} [ 837.803437] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.803775] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.804656] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.804858] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.805155] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.805342] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.805586] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 837.806259] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 838.069694] env[63418]: DEBUG nova.compute.manager [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 838.070786] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a071ca5-f74e-4648-87cf-9c1ae2fa1269 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.082295] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f51e14-c8ab-4ce2-8022-5c6ae8a2534a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.095202] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.319s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.098066] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.132s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.099554] env[63418]: INFO nova.compute.claims [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.123769] env[63418]: WARNING nova.virt.vmwareapi.driver [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 could not be found. [ 838.124041] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 838.124492] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14b7b1e6-825e-435c-a788-eba42abeb217 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.136084] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c17d0ac-4028-486a-8b70-0671b2994184 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.165705] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529d4232-ad53-139c-1c35-049225683b60, 'name': SearchDatastore_Task, 'duration_secs': 0.019036} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.173453] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.173756] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.174538] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 could not be found. [ 838.174788] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 838.175056] env[63418]: INFO nova.compute.manager [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Took 0.11 seconds to destroy the instance on the hypervisor. [ 838.175325] env[63418]: DEBUG oslo.service.loopingcall [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.175569] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89136164-9e79-4ab9-a8dc-1da7e2e51928 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.177637] env[63418]: DEBUG nova.compute.manager [-] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 838.177769] env[63418]: DEBUG nova.network.neutron [-] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.186162] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 838.186162] env[63418]: value = "task-1245036" [ 838.186162] env[63418]: _type = "Task" [ 838.186162] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.195173] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.221700] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245035, 'name': CreateSnapshot_Task, 'duration_secs': 0.414261} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.221985] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 838.222905] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc88e43-eb2d-41da-bde7-bb29d43c18a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.305260] env[63418]: DEBUG nova.objects.base [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 838.305613] env[63418]: DEBUG nova.network.neutron [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.310555] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.411858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6418ba6a-777f-4ca0-982e-380f8f46c2bf tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.121s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.612941] env[63418]: DEBUG oslo_concurrency.lockutils [None req-063fa8f4-2e71-47ff-a16e-709476de4f26 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 42.347s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.614311] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 14.629s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.614311] env[63418]: INFO nova.compute.manager [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Unshelving [ 838.696735] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245036, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.747070] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 838.747879] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0fdaa924-1ef6-4655-8562-1de8ffdfd9fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.760809] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 838.760809] env[63418]: value = "task-1245037" [ 838.760809] env[63418]: _type = "Task" [ 838.760809] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.772653] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245037, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.947530] env[63418]: DEBUG nova.compute.manager [req-f9de80b1-e021-4144-a042-7bfd6d938135 req-0e7108c4-a7a6-4212-b214-e8699747e248 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Received event network-vif-deleted-72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 838.947928] env[63418]: INFO nova.compute.manager [req-f9de80b1-e021-4144-a042-7bfd6d938135 req-0e7108c4-a7a6-4212-b214-e8699747e248 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Neutron deleted interface 72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5; detaching it from the instance and deleting it from the info cache [ 838.952273] env[63418]: DEBUG nova.network.neutron [req-f9de80b1-e021-4144-a042-7bfd6d938135 req-0e7108c4-a7a6-4212-b214-e8699747e248 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.206573] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245036, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517881} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.209502] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.209756] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.210353] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2315e7b-59cb-477c-9070-927ee2c27bf5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.217887] env[63418]: DEBUG nova.network.neutron [-] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.224082] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 839.224082] env[63418]: value = "task-1245038" [ 839.224082] env[63418]: _type = "Task" [ 839.224082] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.241118] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.279830] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245037, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.442746] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d5b9a4-11b4-4dde-a631-7822ba28b770 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.456031] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf1ac95-ba22-42fe-9104-c87f248fa24d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.459187] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8d5a986-d29f-4863-993b-20826289f5b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.494647] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63ef528-e6f9-4938-b6b6-0fcf346d5e02 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.507673] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f6ae20-8f73-4d63-94cb-7193e8f9c040 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.539554] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5c81d2-6655-42c5-a590-a827ac946119 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.546244] env[63418]: DEBUG nova.compute.manager [req-f9de80b1-e021-4144-a042-7bfd6d938135 req-0e7108c4-a7a6-4212-b214-e8699747e248 service nova] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Detach interface failed, port_id=72cfead0-4ef9-4e2f-8ec0-4c11b3a21ca5, reason: Instance 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 839.564160] env[63418]: DEBUG nova.compute.provider_tree [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.625475] env[63418]: DEBUG nova.compute.utils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.721462] env[63418]: INFO nova.compute.manager [-] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Took 1.54 seconds to deallocate network for instance. [ 839.740168] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251608} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.740580] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.743122] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3450090-bedc-4aa3-bf0b-2f161dfaf2d5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.773229] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.773642] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57cc272c-c142-497d-8ce2-329783a89f3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.801987] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245037, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.803544] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 839.803544] env[63418]: value = "task-1245039" [ 839.803544] env[63418]: _type = "Task" [ 839.803544] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.814146] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245039, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.067778] env[63418]: DEBUG nova.scheduler.client.report [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 840.129265] env[63418]: INFO nova.virt.block_device [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Booting with volume ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967 at /dev/sdb [ 840.167206] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94130323-ef28-45f6-972a-eca50dd78d51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.178031] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be0a01b-cc0a-4577-939c-e40c146b9773 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.210949] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6818425a-6dc7-4c06-834b-c42f900db0cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.221682] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e9a7da-a445-45de-9d48-4608a54eb21b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.251613] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91eca91a-73cb-4b4e-9734-9e54cbdf859c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.259373] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8975289d-0858-4386-9c4e-c2ad64fe0870 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.277946] env[63418]: DEBUG nova.virt.block_device [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating existing volume attachment record: 6c5afe76-b23f-4211-a936-38be25c5fc25 {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 840.290082] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245037, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.303239] env[63418]: INFO nova.compute.manager [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Took 0.58 seconds to detach 1 volumes for instance. [ 840.308544] env[63418]: DEBUG nova.compute.manager [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Deleting volume: bef2d2e5-3fa7-48bf-8c27-f4f988f36dce {{(pid=63418) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 840.325131] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.393522] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.393797] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.394319] env[63418]: DEBUG nova.objects.instance [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'flavor' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.575323] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.575977] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 840.579195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.909s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.580731] env[63418]: INFO nova.compute.claims [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.790880] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245037, 'name': CloneVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.818284] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245039, 'name': ReconfigVM_Task, 'duration_secs': 0.686542} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.819319] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Reconfigured VM instance instance-00000047 to attach disk [datastore1] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.819454] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37c37b67-2187-489e-a3c4-8c9b271a8161 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.828057] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 840.828057] env[63418]: value = "task-1245044" [ 840.828057] env[63418]: _type = "Task" [ 840.828057] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.841415] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245044, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.854970] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.022763] env[63418]: DEBUG nova.objects.instance [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'pci_requests' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.088968] env[63418]: DEBUG nova.compute.utils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.090268] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 841.090457] env[63418]: DEBUG nova.network.neutron [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.168853] env[63418]: DEBUG nova.policy [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf4b273bfdc44700be92623cb508d3d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c325dd3ec4e43488c972c9d161e18b1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.291918] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245037, 'name': CloneVM_Task, 'duration_secs': 2.067449} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.292367] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Created linked-clone VM from snapshot [ 841.293497] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7aa6a8e-3432-43bc-9776-6cde192a2692 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.301827] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Uploading image 2ae903eb-37c3-43b7-8fa0-f023b68dbea9 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 841.336512] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 841.336512] env[63418]: value = "vm-268461" [ 841.336512] env[63418]: _type = "VirtualMachine" [ 841.336512] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 841.336963] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-81185985-ea6f-4124-9c15-7cee125e125c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.343014] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245044, 'name': Rename_Task, 'duration_secs': 0.168852} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.343665] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.343920] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edc9fc6a-a7ad-4fbf-98e4-f1ef0503c54d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.350800] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lease: (returnval){ [ 841.350800] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52751fda-9c31-52e9-42e5-0daf0d3567b8" [ 841.350800] env[63418]: _type = "HttpNfcLease" [ 841.350800] env[63418]: } obtained for exporting VM: (result){ [ 841.350800] env[63418]: value = "vm-268461" [ 841.350800] env[63418]: _type = "VirtualMachine" [ 841.350800] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 841.351388] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the lease: (returnval){ [ 841.351388] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52751fda-9c31-52e9-42e5-0daf0d3567b8" [ 841.351388] env[63418]: _type = "HttpNfcLease" [ 841.351388] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 841.355208] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 841.355208] env[63418]: value = "task-1245046" [ 841.355208] env[63418]: _type = "Task" [ 841.355208] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.362655] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 841.362655] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52751fda-9c31-52e9-42e5-0daf0d3567b8" [ 841.362655] env[63418]: _type = "HttpNfcLease" [ 841.362655] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 841.367864] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.513303] env[63418]: DEBUG nova.network.neutron [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Successfully created port: b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.525167] env[63418]: DEBUG nova.objects.base [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 841.525391] env[63418]: DEBUG nova.network.neutron [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.570129] env[63418]: DEBUG nova.policy [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.597143] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 841.860840] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 841.860840] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52751fda-9c31-52e9-42e5-0daf0d3567b8" [ 841.860840] env[63418]: _type = "HttpNfcLease" [ 841.860840] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 841.861581] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 841.861581] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52751fda-9c31-52e9-42e5-0daf0d3567b8" [ 841.861581] env[63418]: _type = "HttpNfcLease" [ 841.861581] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 841.864663] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30338121-003e-4b6e-83c2-a5495b26b825 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.870153] env[63418]: DEBUG oslo_vmware.api [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245046, 'name': PowerOnVM_Task, 'duration_secs': 0.481542} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.870834] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.870935] env[63418]: INFO nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Took 9.76 seconds to spawn the instance on the hypervisor. [ 841.871139] env[63418]: DEBUG nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 841.871899] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3e03be-b2da-4365-8bcb-df366b964671 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.880775] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d57de7-76bb-0783-db1f-1242032d4fc8/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 841.880956] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d57de7-76bb-0783-db1f-1242032d4fc8/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 841.883795] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a2c944-ce76-4421-a459-64b85a8dad63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.949301] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.949564] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.949799] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.950080] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.950352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.953108] env[63418]: DEBUG nova.network.neutron [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Successfully created port: 426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.957347] env[63418]: INFO nova.compute.manager [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Terminating instance [ 841.968346] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9104d8-0c25-43cb-9da5-38de90ade987 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.972458] env[63418]: INFO nova.compute.manager [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Took 34.74 seconds to build instance. [ 842.005376] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bab2255-fb32-4183-a4df-29c3e08460ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.015092] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904de1b0-6e3d-4ae6-9c37-b87e99ff5cca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.029606] env[63418]: DEBUG nova.compute.provider_tree [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.055042] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6a7bf14d-97e5-4325-9275-41c316b86558 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.473339] env[63418]: DEBUG nova.compute.manager [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 842.473593] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.474051] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b20ad44-22fc-4cbd-81a8-a2c3a1ff0bd7 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.250s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.476215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5a2b56-58fc-48c3-a435-e605cdde1b20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.488885] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.488885] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8557ab88-efad-47fe-abe7-80758ca3d66c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.495833] env[63418]: DEBUG oslo_vmware.api [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 842.495833] env[63418]: value = "task-1245047" [ 842.495833] env[63418]: _type = "Task" [ 842.495833] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.507316] env[63418]: DEBUG oslo_vmware.api [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1245047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.532852] env[63418]: DEBUG nova.scheduler.client.report [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 842.611548] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 842.648917] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.649577] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.650201] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.650476] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.654027] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.654027] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.654027] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.654027] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.654027] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.654427] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.654427] env[63418]: DEBUG nova.virt.hardware [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.654427] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496835e6-a3a2-4166-b9a3-06a7c226d553 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.665148] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a850653c-5fd4-45c5-b019-05a69379377c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.006877] env[63418]: DEBUG oslo_vmware.api [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1245047, 'name': PowerOffVM_Task, 'duration_secs': 0.219159} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.007135] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.007312] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.007570] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69f6d951-74af-4b83-a67f-29db968a8440 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.045435] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.045435] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 843.053214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.203s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.053506] env[63418]: DEBUG nova.objects.instance [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lazy-loading 'resources' on Instance uuid 6b81dbe2-aa9e-4561-962b-2af167234b90 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.058559] env[63418]: DEBUG nova.compute.manager [req-b1fde596-485a-4448-909a-84ee19995e8b req-d65e34a8-9a70-40df-9d2d-04edd58cc8d6 service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Received event network-vif-plugged-b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 843.058677] env[63418]: DEBUG oslo_concurrency.lockutils [req-b1fde596-485a-4448-909a-84ee19995e8b req-d65e34a8-9a70-40df-9d2d-04edd58cc8d6 service nova] Acquiring lock "e19a60cd-4ec7-48ad-9042-f19b75353364-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.058885] env[63418]: DEBUG oslo_concurrency.lockutils [req-b1fde596-485a-4448-909a-84ee19995e8b req-d65e34a8-9a70-40df-9d2d-04edd58cc8d6 service nova] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.059169] env[63418]: DEBUG oslo_concurrency.lockutils [req-b1fde596-485a-4448-909a-84ee19995e8b req-d65e34a8-9a70-40df-9d2d-04edd58cc8d6 service nova] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.059532] env[63418]: DEBUG nova.compute.manager [req-b1fde596-485a-4448-909a-84ee19995e8b req-d65e34a8-9a70-40df-9d2d-04edd58cc8d6 service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] No waiting events found dispatching network-vif-plugged-b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 843.060320] env[63418]: WARNING nova.compute.manager [req-b1fde596-485a-4448-909a-84ee19995e8b req-d65e34a8-9a70-40df-9d2d-04edd58cc8d6 service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Received unexpected event network-vif-plugged-b1a1889f-9183-4abc-be7d-fedbe9146896 for instance with vm_state building and task_state spawning. [ 843.097531] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.097804] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.098149] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Deleting the datastore file [datastore2] 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.098598] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f83d7ac-0ac8-418f-8d45-b6563e12b821 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.108861] env[63418]: DEBUG oslo_vmware.api [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for the task: (returnval){ [ 843.108861] env[63418]: value = "task-1245050" [ 843.108861] env[63418]: _type = "Task" [ 843.108861] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.120775] env[63418]: DEBUG oslo_vmware.api [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1245050, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.554787] env[63418]: DEBUG nova.compute.utils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.557102] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 843.557102] env[63418]: DEBUG nova.network.neutron [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.622747] env[63418]: DEBUG oslo_vmware.api [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Task: {'id': task-1245050, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151304} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.627196] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.627442] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.627634] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.627823] env[63418]: INFO nova.compute.manager [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Took 1.15 seconds to destroy the instance on the hypervisor. [ 843.628090] env[63418]: DEBUG oslo.service.loopingcall [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.629447] env[63418]: DEBUG nova.compute.manager [-] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 843.629536] env[63418]: DEBUG nova.network.neutron [-] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 843.649262] env[63418]: DEBUG nova.policy [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea507bad11c3406d880ba47d08a047c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd445600834dd4c7e8022349ee993f3ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.695591] env[63418]: DEBUG nova.network.neutron [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Successfully updated port: 426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.732697] env[63418]: DEBUG nova.network.neutron [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Successfully updated port: b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.869648] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b396f1d1-f30d-423e-836d-73372c338d36 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.881267] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c3d2a8-dca2-46fe-9011-0af819ad3890 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.914694] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907acc72-b0cd-426b-8f08-e8e9784e6aef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.924212] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d705f1f-8686-4197-95cd-21cd87a0ef44 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.939407] env[63418]: DEBUG nova.compute.provider_tree [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.001153] env[63418]: DEBUG nova.compute.manager [req-940278fa-7f1b-4aa7-a629-c75021c3b79d req-405534c3-3661-41a7-a12b-0964a32e680b service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-vif-plugged-426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 844.001153] env[63418]: DEBUG oslo_concurrency.lockutils [req-940278fa-7f1b-4aa7-a629-c75021c3b79d req-405534c3-3661-41a7-a12b-0964a32e680b service nova] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.001984] env[63418]: DEBUG oslo_concurrency.lockutils [req-940278fa-7f1b-4aa7-a629-c75021c3b79d req-405534c3-3661-41a7-a12b-0964a32e680b service nova] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.002170] env[63418]: DEBUG oslo_concurrency.lockutils [req-940278fa-7f1b-4aa7-a629-c75021c3b79d req-405534c3-3661-41a7-a12b-0964a32e680b service nova] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.002457] env[63418]: DEBUG nova.compute.manager [req-940278fa-7f1b-4aa7-a629-c75021c3b79d req-405534c3-3661-41a7-a12b-0964a32e680b service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] No waiting events found dispatching network-vif-plugged-426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 844.002689] env[63418]: WARNING nova.compute.manager [req-940278fa-7f1b-4aa7-a629-c75021c3b79d req-405534c3-3661-41a7-a12b-0964a32e680b service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received unexpected event network-vif-plugged-426ae826-8d40-4981-b7d4-280705c0ac16 for instance with vm_state active and task_state None. [ 844.059994] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 844.159376] env[63418]: DEBUG nova.compute.manager [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Received event network-changed-b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 844.159630] env[63418]: DEBUG nova.compute.manager [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Refreshing instance network info cache due to event network-changed-b1a1889f-9183-4abc-be7d-fedbe9146896. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 844.159884] env[63418]: DEBUG oslo_concurrency.lockutils [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] Acquiring lock "refresh_cache-e19a60cd-4ec7-48ad-9042-f19b75353364" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.160099] env[63418]: DEBUG oslo_concurrency.lockutils [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] Acquired lock "refresh_cache-e19a60cd-4ec7-48ad-9042-f19b75353364" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.160319] env[63418]: DEBUG nova.network.neutron [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Refreshing network info cache for port b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.195329] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.195596] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.195791] env[63418]: DEBUG nova.network.neutron [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.236872] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "refresh_cache-e19a60cd-4ec7-48ad-9042-f19b75353364" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.446038] env[63418]: DEBUG nova.scheduler.client.report [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 844.565733] env[63418]: DEBUG nova.network.neutron [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Successfully created port: 4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.744569] env[63418]: WARNING nova.network.neutron [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] f11c9d94-8e8f-4c7d-b518-f8958080b8fa already exists in list: networks containing: ['f11c9d94-8e8f-4c7d-b518-f8958080b8fa']. ignoring it [ 844.747290] env[63418]: DEBUG nova.network.neutron [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.955975] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.960061] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.732s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.960061] env[63418]: DEBUG nova.objects.instance [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lazy-loading 'resources' on Instance uuid 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.990461] env[63418]: INFO nova.scheduler.client.report [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Deleted allocations for instance 6b81dbe2-aa9e-4561-962b-2af167234b90 [ 845.006526] env[63418]: DEBUG nova.network.neutron [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.074112] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 845.107126] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.107126] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.107126] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.107282] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.107654] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.107654] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.107760] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.107968] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.109399] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.110558] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.110765] env[63418]: DEBUG nova.virt.hardware [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.111708] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf78e1e-3e75-4272-b7ac-10ab3e75c56b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.122101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc87793-b053-411e-97c7-6129a2697837 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.180953] env[63418]: DEBUG nova.network.neutron [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "426ae826-8d40-4981-b7d4-280705c0ac16", "address": "fa:16:3e:1c:8c:cb", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap426ae826-8d", "ovs_interfaceid": "426ae826-8d40-4981-b7d4-280705c0ac16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.502138] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5dfeee59-97d6-47e5-89ad-a2ea0799c4cd tempest-ListServerFiltersTestJSON-1213557408 tempest-ListServerFiltersTestJSON-1213557408-project-member] Lock "6b81dbe2-aa9e-4561-962b-2af167234b90" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.715s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.517185] env[63418]: DEBUG oslo_concurrency.lockutils [req-1124ffe5-98d0-491e-b508-adb043941b90 req-a1453576-4749-43d4-a0f4-a71b55f57afb service nova] Releasing lock "refresh_cache-e19a60cd-4ec7-48ad-9042-f19b75353364" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.517185] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired lock "refresh_cache-e19a60cd-4ec7-48ad-9042-f19b75353364" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.517185] env[63418]: DEBUG nova.network.neutron [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.606440] env[63418]: DEBUG nova.network.neutron [-] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.684381] env[63418]: INFO nova.compute.manager [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Rebuilding instance [ 845.686699] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.687563] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.688272] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.689400] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b72070-2331-4d97-9793-571878364de5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.713592] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.713925] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.714175] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.714424] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.714679] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.714848] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.715139] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.715373] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.715602] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.715821] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.716061] env[63418]: DEBUG nova.virt.hardware [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.723598] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfiguring VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 845.731423] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9f612da-13b5-43e3-8fce-d11cd57d78b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.755019] env[63418]: DEBUG oslo_vmware.api [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 845.755019] env[63418]: value = "task-1245051" [ 845.755019] env[63418]: _type = "Task" [ 845.755019] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.763711] env[63418]: DEBUG oslo_vmware.api [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245051, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.769727] env[63418]: DEBUG nova.compute.manager [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 845.770738] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196c347f-74fd-417f-aac4-ae107972b262 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.813902] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37f3fac-7bd6-4f01-aa0b-534cf797cfe0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.823302] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d7be49-db27-4627-af0b-79e53b3589da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.859599] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08aebbba-b73a-46c8-8bf3-ff1204a3f005 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.869721] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f3ed79-c73c-4c90-bb8a-6843594e02d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.889095] env[63418]: DEBUG nova.compute.provider_tree [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.895938] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.053791] env[63418]: DEBUG nova.network.neutron [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.114473] env[63418]: INFO nova.compute.manager [-] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Took 2.48 seconds to deallocate network for instance. [ 846.264249] env[63418]: DEBUG oslo_vmware.api [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245051, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.300362] env[63418]: DEBUG nova.network.neutron [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Updating instance_info_cache with network_info: [{"id": "b1a1889f-9183-4abc-be7d-fedbe9146896", "address": "fa:16:3e:b3:37:02", "network": {"id": "1c35e426-afbd-4c95-97b6-6e351aab89ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1756487440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c325dd3ec4e43488c972c9d161e18b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1a1889f-91", "ovs_interfaceid": "b1a1889f-9183-4abc-be7d-fedbe9146896", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.391394] env[63418]: DEBUG nova.scheduler.client.report [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 846.621626] env[63418]: DEBUG nova.network.neutron [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Successfully updated port: 4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.624263] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.692984] env[63418]: DEBUG nova.compute.manager [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-changed-426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 846.693217] env[63418]: DEBUG nova.compute.manager [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing instance network info cache due to event network-changed-426ae826-8d40-4981-b7d4-280705c0ac16. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 846.693450] env[63418]: DEBUG oslo_concurrency.lockutils [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.693599] env[63418]: DEBUG oslo_concurrency.lockutils [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.694887] env[63418]: DEBUG nova.network.neutron [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing network info cache for port 426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.765219] env[63418]: DEBUG oslo_vmware.api [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245051, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.789129] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.789343] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a78f216-76ad-4256-ac5d-204861ca51b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.797793] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 846.797793] env[63418]: value = "task-1245052" [ 846.797793] env[63418]: _type = "Task" [ 846.797793] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.806864] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Releasing lock "refresh_cache-e19a60cd-4ec7-48ad-9042-f19b75353364" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.808470] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Instance network_info: |[{"id": "b1a1889f-9183-4abc-be7d-fedbe9146896", "address": "fa:16:3e:b3:37:02", "network": {"id": "1c35e426-afbd-4c95-97b6-6e351aab89ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1756487440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c325dd3ec4e43488c972c9d161e18b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1a1889f-91", "ovs_interfaceid": "b1a1889f-9183-4abc-be7d-fedbe9146896", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 846.808868] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.809436] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:37:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8459aaf-d6a8-46fb-ad14-464ac3104695', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1a1889f-9183-4abc-be7d-fedbe9146896', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.818608] env[63418]: DEBUG oslo.service.loopingcall [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.818746] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.818987] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca12de58-1d70-40a6-b851-30c470757e64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.838498] env[63418]: DEBUG nova.compute.manager [req-be410e4f-bdef-40df-b143-e70849b8eb5b req-7dd56134-3b27-4b05-a4ab-6723fda8e87c service nova] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Received event network-vif-deleted-9b941608-e1b2-4bce-8249-e53c075ac132 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 846.840488] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.840488] env[63418]: value = "task-1245053" [ 846.840488] env[63418]: _type = "Task" [ 846.840488] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.852078] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245053, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.896712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.938s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.900434] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.281s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.900434] env[63418]: DEBUG nova.objects.instance [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lazy-loading 'resources' on Instance uuid f7554ae4-c7a7-4111-a830-10f9029dc074 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.926696] env[63418]: INFO nova.scheduler.client.report [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted allocations for instance 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48 [ 847.091284] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "384b4ae7-ad55-4d44-bd10-3b82534c9703" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.091635] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.126436] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "refresh_cache-1f213c5b-de27-48d6-a1f8-fdf9b77359fd" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.126796] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "refresh_cache-1f213c5b-de27-48d6-a1f8-fdf9b77359fd" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.126957] env[63418]: DEBUG nova.network.neutron [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.265399] env[63418]: DEBUG oslo_vmware.api [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245051, 'name': ReconfigVM_Task, 'duration_secs': 1.333341} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.265958] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.266211] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfigured VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 847.309581] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245052, 'name': PowerOffVM_Task, 'duration_secs': 0.297938} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.309821] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.310074] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 847.310870] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5760b4-1f3e-4434-ae1d-51c80bbb3443 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.321272] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.321465] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4edffd8-d9eb-46b2-b8f7-6569cc222e96 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.358778] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245053, 'name': CreateVM_Task, 'duration_secs': 0.383202} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.358778] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.358778] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.358778] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.359077] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.359400] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-881a1520-e224-453a-ac9c-ebbd58d026c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.365881] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 847.365881] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528bb20c-fa0f-af24-3a4c-257f83b44827" [ 847.365881] env[63418]: _type = "Task" [ 847.365881] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.376077] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528bb20c-fa0f-af24-3a4c-257f83b44827, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.426216] env[63418]: DEBUG nova.network.neutron [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updated VIF entry in instance network info cache for port 426ae826-8d40-4981-b7d4-280705c0ac16. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.426705] env[63418]: DEBUG nova.network.neutron [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "426ae826-8d40-4981-b7d4-280705c0ac16", "address": "fa:16:3e:1c:8c:cb", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap426ae826-8d", "ovs_interfaceid": "426ae826-8d40-4981-b7d4-280705c0ac16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.435438] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f6ccf15-f2ff-48c9-85a3-2b6ee8e7bb14 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.671s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.594382] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 847.642457] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcad70d-d119-4fdc-9445-6efe2fee68d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.652014] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e662cb-8f76-4652-b58f-3b2efec14420 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.688243] env[63418]: DEBUG nova.network.neutron [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.691038] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577024ae-a566-4c3a-956c-8cfe77a53bd8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.700739] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a450e37-fad0-40f1-b4ad-40c3cf32a425 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.716281] env[63418]: DEBUG nova.compute.provider_tree [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.773789] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7cad0b1f-bd20-439a-928c-0292a1ea8596 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.380s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.860354] env[63418]: DEBUG nova.network.neutron [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Updating instance_info_cache with network_info: [{"id": "4a552c2b-4e02-4200-b995-63e7101cf588", "address": "fa:16:3e:60:f3:17", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a552c2b-4e", "ovs_interfaceid": "4a552c2b-4e02-4200-b995-63e7101cf588", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.880892] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528bb20c-fa0f-af24-3a4c-257f83b44827, 'name': SearchDatastore_Task, 'duration_secs': 0.012113} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.881611] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.881911] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.882137] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.882357] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.882505] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.883087] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ec9f749-aeeb-49b0-82bf-8549a44cb0a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.892898] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.893102] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.893849] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af056769-1299-46c6-b0c2-3b97406d65b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.900575] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 847.900575] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5206a776-1d53-ee63-8b41-ace6ffa01f62" [ 847.900575] env[63418]: _type = "Task" [ 847.900575] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.909726] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5206a776-1d53-ee63-8b41-ace6ffa01f62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.932195] env[63418]: DEBUG oslo_concurrency.lockutils [req-46e30019-37ff-43c0-8826-09042a2d4fde req-e2424886-5919-4d30-ad0c-3fb30a0f4807 service nova] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.061441] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Acquiring lock "394744c0-32eb-4c72-8295-5787706d32b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.061781] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Lock "394744c0-32eb-4c72-8295-5787706d32b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.113652] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.219457] env[63418]: DEBUG nova.scheduler.client.report [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 848.363409] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "refresh_cache-1f213c5b-de27-48d6-a1f8-fdf9b77359fd" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.363824] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Instance network_info: |[{"id": "4a552c2b-4e02-4200-b995-63e7101cf588", "address": "fa:16:3e:60:f3:17", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a552c2b-4e", "ovs_interfaceid": "4a552c2b-4e02-4200-b995-63e7101cf588", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 848.364355] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:f3:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a552c2b-4e02-4200-b995-63e7101cf588', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.374427] env[63418]: DEBUG oslo.service.loopingcall [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.375234] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.375604] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-096cc620-f0f8-4422-ac7a-d50a2e4ebc2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.396464] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 848.396521] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 848.396799] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore1] e2dcb8b8-e778-4202-8808-6a8535e2f1b4 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.397518] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a516e40-2766-4ebf-ba6e-b2bcbd78e974 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.408239] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.408239] env[63418]: value = "task-1245055" [ 848.408239] env[63418]: _type = "Task" [ 848.408239] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.408239] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 848.408239] env[63418]: value = "task-1245056" [ 848.408239] env[63418]: _type = "Task" [ 848.408239] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.418596] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5206a776-1d53-ee63-8b41-ace6ffa01f62, 'name': SearchDatastore_Task, 'duration_secs': 0.010302} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.420471] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de07a975-b3ee-4f33-b5fe-395326b06af8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.430039] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.430273] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245055, 'name': CreateVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.434313] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 848.434313] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0e026-a7fa-b2b1-ef84-166039519853" [ 848.434313] env[63418]: _type = "Task" [ 848.434313] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.446198] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0e026-a7fa-b2b1-ef84-166039519853, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.564713] env[63418]: DEBUG nova.compute.manager [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 848.724467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.824s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.728895] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.641s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.729355] env[63418]: DEBUG nova.objects.instance [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lazy-loading 'resources' on Instance uuid be1c134c-f36e-4cc7-b4ef-8f30793fb4df {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.749948] env[63418]: INFO nova.scheduler.client.report [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Deleted allocations for instance f7554ae4-c7a7-4111-a830-10f9029dc074 [ 848.924145] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245055, 'name': CreateVM_Task, 'duration_secs': 0.394653} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.927499] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 848.930352] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20276} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.930352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.930352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.930352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 848.930352] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.930352] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.930780] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.933603] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-025264e1-283f-47c4-91bd-e4fc62181a2d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.943281] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 848.943281] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52171528-2acf-cc22-3ff0-fdb818e1df13" [ 848.943281] env[63418]: _type = "Task" [ 848.943281] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.947656] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0e026-a7fa-b2b1-ef84-166039519853, 'name': SearchDatastore_Task, 'duration_secs': 0.012158} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.952390] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.952815] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e19a60cd-4ec7-48ad-9042-f19b75353364/e19a60cd-4ec7-48ad-9042-f19b75353364.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.956229] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73fce63f-b0ac-49ad-b676-95532d2220ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.967584] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52171528-2acf-cc22-3ff0-fdb818e1df13, 'name': SearchDatastore_Task, 'duration_secs': 0.010225} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.970192] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.970466] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.970715] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.970873] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.971088] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.971411] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 848.971411] env[63418]: value = "task-1245057" [ 848.971411] env[63418]: _type = "Task" [ 848.971411] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.971610] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53132c1c-b456-44f2-9534-494dc9e80126 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.984451] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.984451] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.984540] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.985274] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39b0cdeb-35c6-4864-83ca-679e22136846 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.993502] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 848.993502] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256a35a-873f-4888-8dc6-46f6d2332b20" [ 848.993502] env[63418]: _type = "Task" [ 848.993502] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.006752] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256a35a-873f-4888-8dc6-46f6d2332b20, 'name': SearchDatastore_Task, 'duration_secs': 0.010456} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.010408] env[63418]: DEBUG nova.compute.manager [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Received event network-vif-plugged-4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 849.010670] env[63418]: DEBUG oslo_concurrency.lockutils [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] Acquiring lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.010931] env[63418]: DEBUG oslo_concurrency.lockutils [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.011179] env[63418]: DEBUG oslo_concurrency.lockutils [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.011409] env[63418]: DEBUG nova.compute.manager [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] No waiting events found dispatching network-vif-plugged-4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 849.011638] env[63418]: WARNING nova.compute.manager [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Received unexpected event network-vif-plugged-4a552c2b-4e02-4200-b995-63e7101cf588 for instance with vm_state building and task_state spawning. [ 849.011866] env[63418]: DEBUG nova.compute.manager [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Received event network-changed-4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 849.012073] env[63418]: DEBUG nova.compute.manager [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Refreshing instance network info cache due to event network-changed-4a552c2b-4e02-4200-b995-63e7101cf588. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 849.012279] env[63418]: DEBUG oslo_concurrency.lockutils [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] Acquiring lock "refresh_cache-1f213c5b-de27-48d6-a1f8-fdf9b77359fd" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.012416] env[63418]: DEBUG oslo_concurrency.lockutils [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] Acquired lock "refresh_cache-1f213c5b-de27-48d6-a1f8-fdf9b77359fd" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.012573] env[63418]: DEBUG nova.network.neutron [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Refreshing network info cache for port 4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.013675] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d22f117b-e9a7-44b0-b80a-360925754f85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.024040] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 849.024040] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c41e93-1a19-6610-6dc0-17cab96cf788" [ 849.024040] env[63418]: _type = "Task" [ 849.024040] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.035248] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c41e93-1a19-6610-6dc0-17cab96cf788, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.084413] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.260976] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d062d6c6-44f8-4052-801c-7f946b00c432 tempest-ServerGroupTestJSON-1172547873 tempest-ServerGroupTestJSON-1172547873-project-member] Lock "f7554ae4-c7a7-4111-a830-10f9029dc074" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.148s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.490858] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463206} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.491237] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] e19a60cd-4ec7-48ad-9042-f19b75353364/e19a60cd-4ec7-48ad-9042-f19b75353364.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.491482] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.491784] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4f483a2-91a4-4dff-ad44-0d10bad9739f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.505624] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 849.505624] env[63418]: value = "task-1245058" [ 849.505624] env[63418]: _type = "Task" [ 849.505624] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.523735] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245058, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.526494] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5430477a-609b-475e-97f5-d0677248ebd8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.546288] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff83d06-e041-43bb-8438-94009748e9ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.550357] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c41e93-1a19-6610-6dc0-17cab96cf788, 'name': SearchDatastore_Task, 'duration_secs': 0.011745} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.550687] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.550981] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1f213c5b-de27-48d6-a1f8-fdf9b77359fd/1f213c5b-de27-48d6-a1f8-fdf9b77359fd.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.551704] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbe7f9cc-dd94-47c3-be1e-3d197c07edd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.588029] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77016e9-b54c-49e6-8a7c-5dafeace5df4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.591036] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 849.591036] env[63418]: value = "task-1245059" [ 849.591036] env[63418]: _type = "Task" [ 849.591036] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.598699] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71971e80-fc00-46a8-98c8-b20bd3d4f425 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.606114] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245059, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.617601] env[63418]: DEBUG nova.compute.provider_tree [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.788541] env[63418]: DEBUG nova.network.neutron [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Updated VIF entry in instance network info cache for port 4a552c2b-4e02-4200-b995-63e7101cf588. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.788935] env[63418]: DEBUG nova.network.neutron [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Updating instance_info_cache with network_info: [{"id": "4a552c2b-4e02-4200-b995-63e7101cf588", "address": "fa:16:3e:60:f3:17", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a552c2b-4e", "ovs_interfaceid": "4a552c2b-4e02-4200-b995-63e7101cf588", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.806039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-124ef09c-c406-439c-89c0-b00269da9af4" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.806358] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-124ef09c-c406-439c-89c0-b00269da9af4" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.806845] env[63418]: DEBUG nova.objects.instance [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'flavor' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.986795] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.987085] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.987254] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.987441] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.987594] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.987748] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.987964] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.988551] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.988814] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.989049] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.989288] env[63418]: DEBUG nova.virt.hardware [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.990280] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b13cc5-f0f2-489f-ab90-3dd49c8599f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.000316] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dfc2c2-bc5d-4803-afbc-61774be0ae8a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.017097] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:4e:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf027be2-ee92-4c69-ad07-fe9393549d89', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.025584] env[63418]: DEBUG oslo.service.loopingcall [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.029029] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.029333] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f773ce76-8c23-48b8-9baf-c4a114222c75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.050613] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245058, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075772} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.052123] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.052422] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.052422] env[63418]: value = "task-1245060" [ 850.052422] env[63418]: _type = "Task" [ 850.052422] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.053133] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8718f7-af89-4732-909a-0cdb02b53864 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.070673] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245060, 'name': CreateVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.089804] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] e19a60cd-4ec7-48ad-9042-f19b75353364/e19a60cd-4ec7-48ad-9042-f19b75353364.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.090633] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37b270e0-e536-481f-9d81-47d15356a1a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.118604] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245059, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463643} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.120284] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1f213c5b-de27-48d6-a1f8-fdf9b77359fd/1f213c5b-de27-48d6-a1f8-fdf9b77359fd.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 850.120519] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.121480] env[63418]: DEBUG nova.scheduler.client.report [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 850.125197] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 850.125197] env[63418]: value = "task-1245061" [ 850.125197] env[63418]: _type = "Task" [ 850.125197] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.126042] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7a0e4d8-c2a5-4c84-8ec2-63570ee3007e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.139553] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245061, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.141237] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 850.141237] env[63418]: value = "task-1245062" [ 850.141237] env[63418]: _type = "Task" [ 850.141237] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.152925] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.292371] env[63418]: DEBUG oslo_concurrency.lockutils [req-b8f4f285-b9d8-4a33-a813-18c40c6f0b9f req-56123edb-6cca-4d53-b432-2f1f9eaa91d2 service nova] Releasing lock "refresh_cache-1f213c5b-de27-48d6-a1f8-fdf9b77359fd" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.412905] env[63418]: DEBUG nova.objects.instance [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'pci_requests' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.420049] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d57de7-76bb-0783-db1f-1242032d4fc8/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 850.421531] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c205f6-8a47-4fd1-87a4-5d720105f8ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.433965] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d57de7-76bb-0783-db1f-1242032d4fc8/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 850.434239] env[63418]: ERROR oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d57de7-76bb-0783-db1f-1242032d4fc8/disk-0.vmdk due to incomplete transfer. [ 850.434495] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-48da875d-8de7-4d3b-9500-11a5d5ffe2e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.444187] env[63418]: DEBUG oslo_vmware.rw_handles [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d57de7-76bb-0783-db1f-1242032d4fc8/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 850.444475] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Uploaded image 2ae903eb-37c3-43b7-8fa0-f023b68dbea9 to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 850.447248] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 850.447824] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-773edd98-103e-45a1-a8e9-d8b22d38c207 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.457161] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 850.457161] env[63418]: value = "task-1245063" [ 850.457161] env[63418]: _type = "Task" [ 850.457161] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.468641] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245063, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.571026] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245060, 'name': CreateVM_Task, 'duration_secs': 0.375356} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.571026] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.571026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.571026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.571026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.571026] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd3fd1d2-9bc1-4378-ad22-44f3f077a31c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.575262] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 850.575262] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5291b8fc-80d7-ab63-f0b7-f2d5e434db99" [ 850.575262] env[63418]: _type = "Task" [ 850.575262] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.583840] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5291b8fc-80d7-ab63-f0b7-f2d5e434db99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.629823] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.638020] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.477s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.638020] env[63418]: INFO nova.compute.claims [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.659167] env[63418]: INFO nova.scheduler.client.report [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleted allocations for instance be1c134c-f36e-4cc7-b4ef-8f30793fb4df [ 850.665896] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081976} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.666306] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245061, 'name': ReconfigVM_Task, 'duration_secs': 0.372183} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.668714] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.669086] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Reconfigured VM instance instance-00000048 to attach disk [datastore1] e19a60cd-4ec7-48ad-9042-f19b75353364/e19a60cd-4ec7-48ad-9042-f19b75353364.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.674313] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbce3d67-a2d8-4401-81c4-87a49961b942 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.677250] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cde9cd08-473b-4417-9dba-1c4e14f94f68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.704416] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 1f213c5b-de27-48d6-a1f8-fdf9b77359fd/1f213c5b-de27-48d6-a1f8-fdf9b77359fd.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.706844] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c33f07e-456c-4503-ad91-a1720465fef4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.725209] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 850.725209] env[63418]: value = "task-1245064" [ 850.725209] env[63418]: _type = "Task" [ 850.725209] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.729897] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 850.729897] env[63418]: value = "task-1245065" [ 850.729897] env[63418]: _type = "Task" [ 850.729897] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.737832] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245064, 'name': Rename_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.743341] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245065, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.930505] env[63418]: DEBUG nova.objects.base [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 850.930774] env[63418]: DEBUG nova.network.neutron [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.973772] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245063, 'name': Destroy_Task, 'duration_secs': 0.364722} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.973772] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Destroyed the VM [ 850.973772] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 850.973772] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d7f7f88a-9dfb-4ca6-a5e1-0045b5cbab10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.981048] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 850.981048] env[63418]: value = "task-1245066" [ 850.981048] env[63418]: _type = "Task" [ 850.981048] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.990059] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245066, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.017790] env[63418]: DEBUG nova.policy [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.087561] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5291b8fc-80d7-ab63-f0b7-f2d5e434db99, 'name': SearchDatastore_Task, 'duration_secs': 0.009438} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.088025] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.088868] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.088868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.088868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.088868] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.093135] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed7d1ab1-857a-4a7c-8fe5-5081bd283310 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.099564] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.099740] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.100512] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba7ac679-f81d-4c1d-bc4e-250af16ec360 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.106802] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 851.106802] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a162b6-2db0-a5cc-b5bc-977eeeec8ad5" [ 851.106802] env[63418]: _type = "Task" [ 851.106802] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.115726] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a162b6-2db0-a5cc-b5bc-977eeeec8ad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.182125] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8fc9bb80-9184-4871-9c69-234042675289 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "be1c134c-f36e-4cc7-b4ef-8f30793fb4df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.617s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.235504] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245064, 'name': Rename_Task, 'duration_secs': 0.158369} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.238340] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.238658] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3da14e4c-43a5-4573-8a2b-f7c4f5e8cce0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.254798] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245065, 'name': ReconfigVM_Task, 'duration_secs': 0.305207} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.255932] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 1f213c5b-de27-48d6-a1f8-fdf9b77359fd/1f213c5b-de27-48d6-a1f8-fdf9b77359fd.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.257074] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 851.257074] env[63418]: value = "task-1245067" [ 851.257074] env[63418]: _type = "Task" [ 851.257074] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.257205] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfd65b2f-0293-4958-9f78-4ddbe5259139 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.273426] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 851.273426] env[63418]: value = "task-1245068" [ 851.273426] env[63418]: _type = "Task" [ 851.273426] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.277698] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245067, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.491794] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245066, 'name': RemoveSnapshot_Task, 'duration_secs': 0.392854} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.492154] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 851.492445] env[63418]: DEBUG nova.compute.manager [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 851.493257] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4527aee3-c825-43bb-abc5-13c19dbf38d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.628472] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a162b6-2db0-a5cc-b5bc-977eeeec8ad5, 'name': SearchDatastore_Task, 'duration_secs': 0.009606} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.630127] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4816b221-babc-4921-bb5f-7358b5cdb05d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.639032] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 851.639032] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d29e25-2193-3321-49c8-ed2190757062" [ 851.639032] env[63418]: _type = "Task" [ 851.639032] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.651785] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d29e25-2193-3321-49c8-ed2190757062, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.774926] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245067, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.791767] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245068, 'name': Rename_Task, 'duration_secs': 0.20734} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.794045] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.794509] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efaa04f6-921c-437c-a411-ee8e610bf225 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.806025] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 851.806025] env[63418]: value = "task-1245069" [ 851.806025] env[63418]: _type = "Task" [ 851.806025] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.818649] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.903151] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b0294d-1fe9-4c90-a45e-5d41e3a523cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.911419] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e99bee-3581-499e-89be-d7b801885d3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.949144] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612ad00c-bfc8-48f2-b9b4-694907f1b297 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.958399] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19065e8f-ad68-4d5f-850d-c70b9c63166e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.973933] env[63418]: DEBUG nova.compute.provider_tree [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.008453] env[63418]: INFO nova.compute.manager [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Shelve offloading [ 852.149819] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d29e25-2193-3321-49c8-ed2190757062, 'name': SearchDatastore_Task, 'duration_secs': 0.013746} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.150115] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.150385] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.150652] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3467076c-557e-4d95-87d5-afe56907f45b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.158862] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 852.158862] env[63418]: value = "task-1245070" [ 852.158862] env[63418]: _type = "Task" [ 852.158862] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.169647] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.273262] env[63418]: DEBUG oslo_vmware.api [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245067, 'name': PowerOnVM_Task, 'duration_secs': 0.559705} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.273770] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.274138] env[63418]: INFO nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Took 9.66 seconds to spawn the instance on the hypervisor. [ 852.274682] env[63418]: DEBUG nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 852.275916] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc2780c-0901-40b7-8550-96eebae026d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.319929] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245069, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.479023] env[63418]: DEBUG nova.scheduler.client.report [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 852.516046] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.516046] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e886c4c-9d9f-4ed4-aa76-8d33adb27dd0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.525218] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 852.525218] env[63418]: value = "task-1245071" [ 852.525218] env[63418]: _type = "Task" [ 852.525218] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.537804] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 852.538294] env[63418]: DEBUG nova.compute.manager [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 852.539666] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d6711a-87ee-4ad4-a174-552bd800e71a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.548780] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.548780] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.548780] env[63418]: DEBUG nova.network.neutron [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.597023] env[63418]: DEBUG nova.network.neutron [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Successfully updated port: 124ef09c-c406-439c-89c0-b00269da9af4 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.649382] env[63418]: DEBUG nova.compute.manager [req-40124a8f-9237-44d3-b069-609741a7d7f5 req-7c8b2817-5703-420d-a786-978d9400264a service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-vif-plugged-124ef09c-c406-439c-89c0-b00269da9af4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 852.649720] env[63418]: DEBUG oslo_concurrency.lockutils [req-40124a8f-9237-44d3-b069-609741a7d7f5 req-7c8b2817-5703-420d-a786-978d9400264a service nova] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.650157] env[63418]: DEBUG oslo_concurrency.lockutils [req-40124a8f-9237-44d3-b069-609741a7d7f5 req-7c8b2817-5703-420d-a786-978d9400264a service nova] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.650412] env[63418]: DEBUG oslo_concurrency.lockutils [req-40124a8f-9237-44d3-b069-609741a7d7f5 req-7c8b2817-5703-420d-a786-978d9400264a service nova] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.650644] env[63418]: DEBUG nova.compute.manager [req-40124a8f-9237-44d3-b069-609741a7d7f5 req-7c8b2817-5703-420d-a786-978d9400264a service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] No waiting events found dispatching network-vif-plugged-124ef09c-c406-439c-89c0-b00269da9af4 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 852.650868] env[63418]: WARNING nova.compute.manager [req-40124a8f-9237-44d3-b069-609741a7d7f5 req-7c8b2817-5703-420d-a786-978d9400264a service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received unexpected event network-vif-plugged-124ef09c-c406-439c-89c0-b00269da9af4 for instance with vm_state active and task_state None. [ 852.672246] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245070, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.804039] env[63418]: INFO nova.compute.manager [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Took 30.86 seconds to build instance. [ 852.814216] env[63418]: DEBUG oslo_vmware.api [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245069, 'name': PowerOnVM_Task, 'duration_secs': 0.733987} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.814598] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.814780] env[63418]: INFO nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Took 7.74 seconds to spawn the instance on the hypervisor. [ 852.814934] env[63418]: DEBUG nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 852.815834] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820a13b6-8aca-417e-8c4e-83660c1b6bdd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.984327] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.984912] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 852.987637] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.183s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.987836] env[63418]: DEBUG nova.objects.instance [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lazy-loading 'resources' on Instance uuid fa4c0315-1ef5-4491-ab18-d49563b778fb {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.098608] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.098825] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.098940] env[63418]: DEBUG nova.network.neutron [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.172233] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596047} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.172539] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.172813] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.173124] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14a95cea-aea3-429e-a042-075b79cb7cd3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.181020] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 853.181020] env[63418]: value = "task-1245072" [ 853.181020] env[63418]: _type = "Task" [ 853.181020] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.195135] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.306839] env[63418]: DEBUG nova.network.neutron [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updating instance_info_cache with network_info: [{"id": "0591996d-662a-4dac-80d1-c81781f6e17d", "address": "fa:16:3e:d5:0b:db", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0591996d-66", "ovs_interfaceid": "0591996d-662a-4dac-80d1-c81781f6e17d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.309162] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a786bafd-6f5f-4b76-89ca-8d7226e17ff0 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.380s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.333198] env[63418]: INFO nova.compute.manager [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Took 29.69 seconds to build instance. [ 853.493110] env[63418]: DEBUG nova.compute.utils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.495826] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 853.496284] env[63418]: DEBUG nova.network.neutron [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.557358] env[63418]: DEBUG nova.policy [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d29f8428db741c0a6c9f6f6dfdf50d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0462b212fa4449c2a6f98cec2f186f51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.692071] env[63418]: WARNING nova.network.neutron [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] f11c9d94-8e8f-4c7d-b518-f8958080b8fa already exists in list: networks containing: ['f11c9d94-8e8f-4c7d-b518-f8958080b8fa']. ignoring it [ 853.692333] env[63418]: WARNING nova.network.neutron [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] f11c9d94-8e8f-4c7d-b518-f8958080b8fa already exists in list: networks containing: ['f11c9d94-8e8f-4c7d-b518-f8958080b8fa']. ignoring it [ 853.702678] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067312} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.705377] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.706782] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dcf2a2-ba2b-42ab-9c19-1cca75b118c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.730553] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.731781] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce67ce74-9c90-46ea-bcb8-fe9d6da0bd28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.749825] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a88c82-38cc-4969-b56d-f18da766a75b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.759165] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c66e6b5-779d-47b0-a41c-6ff76b87d243 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.762360] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 853.762360] env[63418]: value = "task-1245073" [ 853.762360] env[63418]: _type = "Task" [ 853.762360] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.793666] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f2ee80-591b-4b44-9477-ae119f3b7b31 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.800384] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.805561] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94de75d3-bb6b-4339-b4b0-fe8be5d0f157 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.811478] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.554741] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d5f3814-5793-4eab-9395-48422cd5c957 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.919s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.555351] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 854.558575] env[63418]: DEBUG nova.compute.provider_tree [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.560221] env[63418]: DEBUG nova.network.neutron [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Successfully created port: c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.574287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "5ab04ddd-4cd6-4330-a9d7-081290c27586" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.574524] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.582208] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245073, 'name': ReconfigVM_Task, 'duration_secs': 0.710944} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.582478] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Reconfigured VM instance instance-00000047 to attach disk [datastore2] e2dcb8b8-e778-4202-8808-6a8535e2f1b4/e2dcb8b8-e778-4202-8808-6a8535e2f1b4.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.583312] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bdb787d9-4486-4df1-890a-b9ad50145a6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.592023] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 854.592023] env[63418]: value = "task-1245074" [ 854.592023] env[63418]: _type = "Task" [ 854.592023] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.603224] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245074, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.005461] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.005461] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd772fa-57ae-4c85-a446-874357836751 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.015274] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 855.015274] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-239a537c-b2c3-4221-84e1-e1c74eb31e97 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.077021] env[63418]: DEBUG nova.scheduler.client.report [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 855.078245] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 855.103027] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 855.103027] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 855.103027] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleting the datastore file [datastore2] 61af37a5-8c18-47a7-817a-eb3b332b3725 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.103027] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67ae5235-ebaa-47f3-9b45-a767058a16b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.108272] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245074, 'name': Rename_Task, 'duration_secs': 0.381633} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.109068] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.109488] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf7fd370-8bcc-41e2-b794-d856b3b989ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.114169] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 855.114169] env[63418]: value = "task-1245076" [ 855.114169] env[63418]: _type = "Task" [ 855.114169] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.119078] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 855.119078] env[63418]: value = "task-1245077" [ 855.119078] env[63418]: _type = "Task" [ 855.119078] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.126159] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.134359] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245077, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.572833] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 855.584641] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.597s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.593623] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.569s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.593835] env[63418]: DEBUG nova.objects.instance [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lazy-loading 'resources' on Instance uuid 849552d4-0bb4-48ee-af7b-390183bd2189 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.616343] env[63418]: DEBUG nova.network.neutron [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "426ae826-8d40-4981-b7d4-280705c0ac16", "address": "fa:16:3e:1c:8c:cb", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap426ae826-8d", "ovs_interfaceid": "426ae826-8d40-4981-b7d4-280705c0ac16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "124ef09c-c406-439c-89c0-b00269da9af4", "address": "fa:16:3e:ed:1d:47", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124ef09c-c4", "ovs_interfaceid": "124ef09c-c406-439c-89c0-b00269da9af4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.620181] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.620181] env[63418]: INFO nova.scheduler.client.report [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Deleted allocations for instance fa4c0315-1ef5-4491-ab18-d49563b778fb [ 855.630901] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.631191] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.631349] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.631529] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.631672] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.631814] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.632051] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.632618] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.632827] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.632963] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.633169] env[63418]: DEBUG nova.virt.hardware [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.639036] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a519a4ea-f3c5-4bc2-8bdc-3e035d862ce1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.652929] env[63418]: DEBUG oslo_vmware.api [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202281} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.658326] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.658582] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.658768] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.660807] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245077, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.662667] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0a7c1b-1a1b-4210-a30c-971f614b43d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.686530] env[63418]: INFO nova.scheduler.client.report [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocations for instance 61af37a5-8c18-47a7-817a-eb3b332b3725 [ 855.907173] env[63418]: DEBUG nova.compute.manager [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-changed-124ef09c-c406-439c-89c0-b00269da9af4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 855.907392] env[63418]: DEBUG nova.compute.manager [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing instance network info cache due to event network-changed-124ef09c-c406-439c-89c0-b00269da9af4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 855.907796] env[63418]: DEBUG oslo_concurrency.lockutils [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.121854] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a47435e-af0b-4e77-a6c7-fcb0e070fa19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.134622] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Suspending the VM {{(pid=63418) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 856.139286] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-803e0288-84b6-4252-af2d-c53d246f6a94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.141824] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.142471] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.142629] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.147041] env[63418]: DEBUG oslo_concurrency.lockutils [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.147041] env[63418]: DEBUG nova.network.neutron [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Refreshing network info cache for port 124ef09c-c406-439c-89c0-b00269da9af4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.147940] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05de770f-562d-48bb-b73b-8920d22741a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.150681] env[63418]: DEBUG oslo_vmware.api [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245077, 'name': PowerOnVM_Task, 'duration_secs': 0.949106} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.153645] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.154057] env[63418]: DEBUG nova.compute.manager [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 856.154813] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c20b2d71-ae7e-4fc1-a99d-3d8668fa23ba tempest-SecurityGroupsTestJSON-1438776888 tempest-SecurityGroupsTestJSON-1438776888-project-member] Lock "fa4c0315-1ef5-4491-ab18-d49563b778fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.896s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.158837] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099affb6-62b7-4d82-bf65-8842080144ba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.175107] env[63418]: DEBUG oslo_vmware.api [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 856.175107] env[63418]: value = "task-1245078" [ 856.175107] env[63418]: _type = "Task" [ 856.175107] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.176706] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.176928] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.177098] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.177281] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.177498] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.177565] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.177769] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.177927] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.178113] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.178278] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.179279] env[63418]: DEBUG nova.virt.hardware [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.184926] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfiguring VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 856.188494] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cd1d1c1-4655-4877-b847-e3ee2499b790 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.214272] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.222547] env[63418]: DEBUG oslo_vmware.api [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245078, 'name': SuspendVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.227188] env[63418]: DEBUG oslo_vmware.api [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 856.227188] env[63418]: value = "task-1245079" [ 856.227188] env[63418]: _type = "Task" [ 856.227188] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.236465] env[63418]: DEBUG oslo_vmware.api [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245079, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.352352] env[63418]: DEBUG nova.network.neutron [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Successfully updated port: c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.393280] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a70e15-c86d-4975-a191-edfd9f913835 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.402285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032f1d9f-7114-489a-a392-03cfb5411aba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.442120] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d72311-7c5c-4151-a266-d05b11963a99 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.452861] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af92ae05-ab6a-4341-a506-52dfbc75d3fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.469165] env[63418]: DEBUG nova.compute.provider_tree [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.698941] env[63418]: DEBUG oslo_vmware.api [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245078, 'name': SuspendVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.726666] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.741076] env[63418]: DEBUG oslo_vmware.api [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.855680] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "refresh_cache-dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.855836] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "refresh_cache-dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.856149] env[63418]: DEBUG nova.network.neutron [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.946071] env[63418]: DEBUG nova.compute.manager [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 856.947547] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e45b6f-1319-42ae-a09b-7a426415235d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.974140] env[63418]: DEBUG nova.scheduler.client.report [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 857.012238] env[63418]: DEBUG nova.network.neutron [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updated VIF entry in instance network info cache for port 124ef09c-c406-439c-89c0-b00269da9af4. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.012776] env[63418]: DEBUG nova.network.neutron [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "426ae826-8d40-4981-b7d4-280705c0ac16", "address": "fa:16:3e:1c:8c:cb", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap426ae826-8d", "ovs_interfaceid": "426ae826-8d40-4981-b7d4-280705c0ac16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "124ef09c-c406-439c-89c0-b00269da9af4", "address": "fa:16:3e:ed:1d:47", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124ef09c-c4", "ovs_interfaceid": "124ef09c-c406-439c-89c0-b00269da9af4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.200489] env[63418]: DEBUG oslo_vmware.api [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245078, 'name': SuspendVM_Task, 'duration_secs': 1.040706} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.200859] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Suspended the VM {{(pid=63418) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 857.200859] env[63418]: DEBUG nova.compute.manager [None req-2054a713-199d-40f4-9246-ead2483df3f6 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.202176] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f3e234-af60-45f2-8684-e360d58148cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.238758] env[63418]: DEBUG oslo_vmware.api [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245079, 'name': ReconfigVM_Task, 'duration_secs': 0.724608} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.239396] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.239658] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfigured VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 857.420439] env[63418]: DEBUG nova.network.neutron [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.462531] env[63418]: INFO nova.compute.manager [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] instance snapshotting [ 857.463193] env[63418]: DEBUG nova.objects.instance [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'flavor' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.481167] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.888s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.485811] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.864s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.510234] env[63418]: INFO nova.scheduler.client.report [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Deleted allocations for instance 849552d4-0bb4-48ee-af7b-390183bd2189 [ 857.522148] env[63418]: DEBUG oslo_concurrency.lockutils [req-83bd8c0a-2474-49bb-b6ef-5823a544095e req-539f20a5-d5c7-4746-b296-d5e08d71b9f3 service nova] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.702037] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48da5f7e-b070-434f-91ae-9527666ec2e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.716541] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b56d6cc-0d23-4100-b560-4210705fb57b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.772248] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5e6028c0-b162-4da7-953c-eb99d7e4a0dd tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-124ef09c-c406-439c-89c0-b00269da9af4" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.966s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.779255] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27aa0090-c4b7-4a9a-9308-3f0fe3b70f88 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.793473] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcb911c-fdf1-44a9-86ac-479b632d89fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.816049] env[63418]: DEBUG nova.compute.provider_tree [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.922118] env[63418]: DEBUG nova.network.neutron [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Updating instance_info_cache with network_info: [{"id": "c3c91249-1c73-4686-baaa-8a7fe085b98d", "address": "fa:16:3e:98:f3:8b", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3c91249-1c", "ovs_interfaceid": "c3c91249-1c73-4686-baaa-8a7fe085b98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.940153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "61af37a5-8c18-47a7-817a-eb3b332b3725" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.975357] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80990220-6c41-4cca-8e09-818cd4808b4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.993720] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d2ddd6-701e-42af-8074-7ff5114db147 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.024052] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c9b57acf-2dd2-4e54-99cf-cab03a6bc5a2 tempest-ServerTagsTestJSON-1834545877 tempest-ServerTagsTestJSON-1834545877-project-member] Lock "849552d4-0bb4-48ee-af7b-390183bd2189" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.738s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.322158] env[63418]: DEBUG nova.scheduler.client.report [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 858.425638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "refresh_cache-dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.425993] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Instance network_info: |[{"id": "c3c91249-1c73-4686-baaa-8a7fe085b98d", "address": "fa:16:3e:98:f3:8b", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3c91249-1c", "ovs_interfaceid": "c3c91249-1c73-4686-baaa-8a7fe085b98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 858.426721] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:f3:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3c91249-1c73-4686-baaa-8a7fe085b98d', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.437862] env[63418]: DEBUG oslo.service.loopingcall [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.438122] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.439755] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f64f2c92-7c71-480a-badb-f20db874944f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.470578] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.470578] env[63418]: value = "task-1245080" [ 858.470578] env[63418]: _type = "Task" [ 858.470578] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.483495] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245080, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.506074] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 858.509022] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4f17aaa0-ebd1-450b-b384-aa29bcab7e20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.517649] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 858.517649] env[63418]: value = "task-1245081" [ 858.517649] env[63418]: _type = "Task" [ 858.517649] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.529468] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245081, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.586238] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "e19a60cd-4ec7-48ad-9042-f19b75353364" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.586605] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.586857] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "e19a60cd-4ec7-48ad-9042-f19b75353364-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.587126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.587342] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.592526] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Received event network-vif-unplugged-0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 858.592526] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Acquiring lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.592526] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.592696] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.592866] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] No waiting events found dispatching network-vif-unplugged-0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.594425] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Received event network-vif-unplugged-0591996d-662a-4dac-80d1-c81781f6e17d for instance with task_state deleting. {{(pid=63418) _process_instance_event /opt/stack/nova/nova/compute/manager.py:11259}} [ 858.594425] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Received event network-changed-0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 858.594425] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Refreshing instance network info cache due to event network-changed-0591996d-662a-4dac-80d1-c81781f6e17d. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 858.594425] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Acquiring lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.594425] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Acquired lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.594425] env[63418]: DEBUG nova.network.neutron [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Refreshing network info cache for port 0591996d-662a-4dac-80d1-c81781f6e17d {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.599068] env[63418]: INFO nova.compute.manager [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Terminating instance [ 858.827423] env[63418]: DEBUG oslo_concurrency.lockutils [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.342s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.827763] env[63418]: INFO nova.compute.manager [None req-fccb38d6-b243-4660-aa0c-9a5ec3666bc5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Successfully reverted task state from rebuilding on failure for instance. [ 858.834509] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.524s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.835050] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.838028] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 858.838028] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.981s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.838028] env[63418]: DEBUG nova.objects.instance [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lazy-loading 'resources' on Instance uuid 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.838483] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00436b3-6e56-4d11-a858-d02010693441 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.849222] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b531d09-2d07-4555-93ad-d3263b881dab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.869653] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f40ee14-5a69-4ad0-9f81-c300493e38d6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.880761] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354198db-cf9a-4aaf-b7bb-7a4da8279b6c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.918891] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180129MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 858.918891] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.983825] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245080, 'name': CreateVM_Task, 'duration_secs': 0.428773} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.984330] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.985607] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.986194] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.986693] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.987189] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2612aa8-4fec-4cb2-9b34-c46f5af23d66 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.996506] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 858.996506] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52021216-4887-8a1c-432b-9cdcc0144ce9" [ 858.996506] env[63418]: _type = "Task" [ 858.996506] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.006650] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52021216-4887-8a1c-432b-9cdcc0144ce9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.031421] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245081, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.102534] env[63418]: DEBUG nova.compute.manager [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 859.102534] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.107135] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63133faa-6220-4522-90ff-4875f84479d5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.117321] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.117903] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c439079-2c98-4123-a26f-6abf4690ecfe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.128124] env[63418]: DEBUG oslo_vmware.api [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 859.128124] env[63418]: value = "task-1245082" [ 859.128124] env[63418]: _type = "Task" [ 859.128124] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.138011] env[63418]: DEBUG oslo_vmware.api [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.388491] env[63418]: DEBUG nova.network.neutron [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updated VIF entry in instance network info cache for port 0591996d-662a-4dac-80d1-c81781f6e17d. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.388950] env[63418]: DEBUG nova.network.neutron [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updating instance_info_cache with network_info: [{"id": "0591996d-662a-4dac-80d1-c81781f6e17d", "address": "fa:16:3e:d5:0b:db", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": null, "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0591996d-66", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.510331] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52021216-4887-8a1c-432b-9cdcc0144ce9, 'name': SearchDatastore_Task, 'duration_secs': 0.013346} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.512340] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.512340] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.512340] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.512340] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.512938] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.516892] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0875c897-5694-40f4-9eed-736cc5075c73 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.530520] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245081, 'name': CreateSnapshot_Task, 'duration_secs': 0.593274} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.531979] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 859.532454] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.532767] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.534093] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af023bf-a577-452f-a83e-43e03173edc8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.537466] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75f7c602-6bb3-4e78-a8a7-ce92bbaba90a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.562251] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 859.562251] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52548c2d-1df4-3327-fdf9-d924d78b66d8" [ 859.562251] env[63418]: _type = "Task" [ 859.562251] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.577303] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52548c2d-1df4-3327-fdf9-d924d78b66d8, 'name': SearchDatastore_Task, 'duration_secs': 0.021052} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.577567] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b9cd909-62aa-44b4-963d-c4fd558f4ffe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.585661] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 859.585661] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ffd8d8-1676-1140-0520-5e6f57debf2e" [ 859.585661] env[63418]: _type = "Task" [ 859.585661] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.596458] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ffd8d8-1676-1140-0520-5e6f57debf2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.598283] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a95e9cf-b081-45a1-9a11-ee3baafd47cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.609029] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5618bb05-092a-4ec4-8c0d-8c906ce8cfbe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.649141] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0e3015-67ef-4064-b8d1-3bfb073c4adf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.653085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.653085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.653085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.653418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.653418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.656396] env[63418]: INFO nova.compute.manager [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Terminating instance [ 859.663582] env[63418]: DEBUG oslo_vmware.api [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245082, 'name': PowerOffVM_Task, 'duration_secs': 0.395584} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.665017] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7976fd3-7033-4c18-913a-0d09df8fcbd2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.670046] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.670046] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.670543] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-835f8e45-8c62-407f-9962-7f5d265d517b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.683093] env[63418]: DEBUG nova.compute.provider_tree [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.751864] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.752319] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.752757] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Deleting the datastore file [datastore1] e19a60cd-4ec7-48ad-9042-f19b75353364 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.753049] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-657565b8-5f8d-461a-a683-111f897f4ed8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.762508] env[63418]: DEBUG oslo_vmware.api [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 859.762508] env[63418]: value = "task-1245084" [ 859.762508] env[63418]: _type = "Task" [ 859.762508] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.771399] env[63418]: DEBUG oslo_vmware.api [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245084, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.891986] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Releasing lock "refresh_cache-61af37a5-8c18-47a7-817a-eb3b332b3725" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.892831] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Received event network-vif-plugged-c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 859.892831] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Acquiring lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.893019] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.894212] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.894212] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] No waiting events found dispatching network-vif-plugged-c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 859.894212] env[63418]: WARNING nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Received unexpected event network-vif-plugged-c3c91249-1c73-4686-baaa-8a7fe085b98d for instance with vm_state building and task_state spawning. [ 859.894212] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Received event network-changed-c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 859.894212] env[63418]: DEBUG nova.compute.manager [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Refreshing instance network info cache due to event network-changed-c3c91249-1c73-4686-baaa-8a7fe085b98d. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 859.894212] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Acquiring lock "refresh_cache-dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.894454] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Acquired lock "refresh_cache-dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.894627] env[63418]: DEBUG nova.network.neutron [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Refreshing network info cache for port c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.070436] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 860.070789] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c92aada5-b9c0-4486-bdf5-3c159032fa39 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.082477] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 860.082477] env[63418]: value = "task-1245085" [ 860.082477] env[63418]: _type = "Task" [ 860.082477] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.097145] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ffd8d8-1676-1140-0520-5e6f57debf2e, 'name': SearchDatastore_Task, 'duration_secs': 0.011654} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.100528] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.101245] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] dcc94934-d8d2-46d9-85fd-e17c11a7a4d7/dcc94934-d8d2-46d9-85fd-e17c11a7a4d7.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.101245] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245085, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.101405] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b2d9d5e-88b8-48c8-a584-af1053500130 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.109711] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 860.109711] env[63418]: value = "task-1245086" [ 860.109711] env[63418]: _type = "Task" [ 860.109711] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.124140] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245086, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.165496] env[63418]: DEBUG nova.compute.manager [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 860.165759] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.166694] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99f50c5-87a7-446a-8ddb-89f791147939 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.176765] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.177175] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-637799c9-0768-41a3-80dd-d5d6768d5861 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.186381] env[63418]: DEBUG nova.scheduler.client.report [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 860.190978] env[63418]: DEBUG oslo_vmware.api [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 860.190978] env[63418]: value = "task-1245087" [ 860.190978] env[63418]: _type = "Task" [ 860.190978] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.203029] env[63418]: DEBUG oslo_vmware.api [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.275413] env[63418]: DEBUG oslo_vmware.api [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245084, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.397902} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.275929] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.276199] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.276487] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.276883] env[63418]: INFO nova.compute.manager [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Took 1.17 seconds to destroy the instance on the hypervisor. [ 860.277217] env[63418]: DEBUG oslo.service.loopingcall [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.278043] env[63418]: DEBUG nova.compute.manager [-] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 860.278043] env[63418]: DEBUG nova.network.neutron [-] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.451839] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "586f51a0-90de-4f44-ac3e-758ceda9e316" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.452331] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.593371] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245085, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.623712] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245086, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.629521] env[63418]: DEBUG nova.compute.manager [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 860.629521] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28451447-5596-4c41-a7b5-1eab0b05fb4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.692818] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.857s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.695708] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.800s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.695844] env[63418]: DEBUG nova.objects.instance [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'pci_requests' on Instance uuid d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.706884] env[63418]: DEBUG oslo_vmware.api [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245087, 'name': PowerOffVM_Task, 'duration_secs': 0.206657} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.713019] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.713019] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.714688] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10011041-b0a9-4668-b625-e16fba93c669 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.786633] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.786989] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.787307] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore2] e2dcb8b8-e778-4202-8808-6a8535e2f1b4 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.787619] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b0d4658-9331-49b2-86fa-97e4f244a8e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.794642] env[63418]: DEBUG nova.compute.manager [req-d770f934-f5d8-4561-b4cb-b1933a2d2295 req-f8590897-9529-4c24-8410-b61b428bbbda service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Received event network-vif-deleted-b1a1889f-9183-4abc-be7d-fedbe9146896 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 860.794843] env[63418]: INFO nova.compute.manager [req-d770f934-f5d8-4561-b4cb-b1933a2d2295 req-f8590897-9529-4c24-8410-b61b428bbbda service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Neutron deleted interface b1a1889f-9183-4abc-be7d-fedbe9146896; detaching it from the instance and deleting it from the info cache [ 860.795193] env[63418]: DEBUG nova.network.neutron [req-d770f934-f5d8-4561-b4cb-b1933a2d2295 req-f8590897-9529-4c24-8410-b61b428bbbda service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.799823] env[63418]: DEBUG oslo_vmware.api [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 860.799823] env[63418]: value = "task-1245089" [ 860.799823] env[63418]: _type = "Task" [ 860.799823] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.809598] env[63418]: DEBUG oslo_vmware.api [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.854552] env[63418]: DEBUG nova.network.neutron [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Updated VIF entry in instance network info cache for port c3c91249-1c73-4686-baaa-8a7fe085b98d. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.854552] env[63418]: DEBUG nova.network.neutron [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Updating instance_info_cache with network_info: [{"id": "c3c91249-1c73-4686-baaa-8a7fe085b98d", "address": "fa:16:3e:98:f3:8b", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3c91249-1c", "ovs_interfaceid": "c3c91249-1c73-4686-baaa-8a7fe085b98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.888179] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.889687] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.958022] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 861.097190] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245085, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.122170] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245086, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521995} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.122745] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] dcc94934-d8d2-46d9-85fd-e17c11a7a4d7/dcc94934-d8d2-46d9-85fd-e17c11a7a4d7.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.123315] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.124204] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1460ba7-633f-4b60-86cb-783babffe16b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.132741] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 861.132741] env[63418]: value = "task-1245090" [ 861.132741] env[63418]: _type = "Task" [ 861.132741] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.145165] env[63418]: INFO nova.compute.manager [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] instance snapshotting [ 861.145165] env[63418]: WARNING nova.compute.manager [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 861.150474] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245090, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.150474] env[63418]: DEBUG nova.network.neutron [-] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.150474] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16f72e9-1a51-4a6e-a3f9-879d77849d2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.171410] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bfe0ad-16a7-43be-8460-77ab2a44ecc9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.202935] env[63418]: DEBUG nova.objects.instance [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'numa_topology' on Instance uuid d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.220050] env[63418]: DEBUG oslo_concurrency.lockutils [None req-93dad61c-f12e-44b2-9385-da4a7d93eba5 tempest-ServerActionsV293TestJSON-702325390 tempest-ServerActionsV293TestJSON-702325390-project-member] Lock "1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.662s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.252491] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-426ae826-8d40-4981-b7d4-280705c0ac16" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.252771] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-426ae826-8d40-4981-b7d4-280705c0ac16" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.301305] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45c621e8-8881-41a2-8117-a6aad0606340 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.312358] env[63418]: DEBUG oslo_vmware.api [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245089, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13399} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.313512] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.313713] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.313895] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.314109] env[63418]: INFO nova.compute.manager [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 861.314373] env[63418]: DEBUG oslo.service.loopingcall [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.314630] env[63418]: DEBUG nova.compute.manager [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 861.314721] env[63418]: DEBUG nova.network.neutron [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 861.318867] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a776c7-2454-47f2-9c61-cd661702b717 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.347409] env[63418]: DEBUG nova.compute.manager [req-d770f934-f5d8-4561-b4cb-b1933a2d2295 req-f8590897-9529-4c24-8410-b61b428bbbda service nova] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Detach interface failed, port_id=b1a1889f-9183-4abc-be7d-fedbe9146896, reason: Instance e19a60cd-4ec7-48ad-9042-f19b75353364 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 861.356329] env[63418]: DEBUG oslo_concurrency.lockutils [req-f952dc23-5900-4b89-bd96-500dea2b9e11 req-796da1c4-00af-4940-a3d8-cfe0a000d540 service nova] Releasing lock "refresh_cache-dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.394409] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 861.479801] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.593964] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245085, 'name': CloneVM_Task, 'duration_secs': 1.492361} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.594907] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created linked-clone VM from snapshot [ 861.595715] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30ae13f-b5a3-4f6c-a42a-f7fa3e7709b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.604250] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Uploading image 218c0554-98fc-478a-bef5-e1a4088a8355 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 861.630619] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 861.630619] env[63418]: value = "vm-268469" [ 861.630619] env[63418]: _type = "VirtualMachine" [ 861.630619] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 861.630923] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-105beebe-806a-49ad-9d4f-15bb468b6630 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.640603] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease: (returnval){ [ 861.640603] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e78f74-d938-577e-0daa-3b7fb7d3cc8d" [ 861.640603] env[63418]: _type = "HttpNfcLease" [ 861.640603] env[63418]: } obtained for exporting VM: (result){ [ 861.640603] env[63418]: value = "vm-268469" [ 861.640603] env[63418]: _type = "VirtualMachine" [ 861.640603] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 861.640890] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the lease: (returnval){ [ 861.640890] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e78f74-d938-577e-0daa-3b7fb7d3cc8d" [ 861.640890] env[63418]: _type = "HttpNfcLease" [ 861.640890] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 861.649372] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245090, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069841} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.649866] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.650711] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf08121f-a591-4401-82d5-c6997da5a93d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.657346] env[63418]: INFO nova.compute.manager [-] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Took 1.38 seconds to deallocate network for instance. [ 861.657600] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 861.657600] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e78f74-d938-577e-0daa-3b7fb7d3cc8d" [ 861.657600] env[63418]: _type = "HttpNfcLease" [ 861.657600] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 861.659708] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 861.659708] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e78f74-d938-577e-0daa-3b7fb7d3cc8d" [ 861.659708] env[63418]: _type = "HttpNfcLease" [ 861.659708] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 861.663204] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5c1307-bd72-4d25-bf1f-6e3428cdd68f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.684093] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] dcc94934-d8d2-46d9-85fd-e17c11a7a4d7/dcc94934-d8d2-46d9-85fd-e17c11a7a4d7.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.686305] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 861.686305] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90fa200e-c8a1-4d31-bb64-81c62c03fa39 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.704229] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d9591c1d-145a-4f9c-9454-31c17be8a4da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.706833] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237e16e-5036-4b50-804a-34917498bb1c/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 861.707017] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237e16e-5036-4b50-804a-34917498bb1c/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 861.708425] env[63418]: INFO nova.compute.claims [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.774970] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.775266] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.782057] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06e2c06-6fd2-410d-9613-ed45f76021e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.784909] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 861.784909] env[63418]: value = "task-1245092" [ 861.784909] env[63418]: _type = "Task" [ 861.784909] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.785195] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 861.785195] env[63418]: value = "task-1245093" [ 861.785195] env[63418]: _type = "Task" [ 861.785195] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.806582] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b28c98-688e-47df-8034-2b8f0590dfe8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.817466] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245093, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.817466] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245092, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.820306] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-124733d3-223f-48ce-b107-8bc4295c048d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.845703] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfiguring VM to detach interface {{(pid=63418) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 861.848965] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20a1ff7e-4da1-4de4-9f21-e1a7b0314e65 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.871263] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 861.871263] env[63418]: value = "task-1245094" [ 861.871263] env[63418]: _type = "Task" [ 861.871263] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.884909] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.928975] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.005997] env[63418]: DEBUG nova.compute.manager [req-d7a924cd-48c4-4aaa-b119-2d3a79c22e4e req-238bbb8b-c20e-403e-b61f-5ce66d61d95f service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Received event network-vif-deleted-bf027be2-ee92-4c69-ad07-fe9393549d89 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 862.006283] env[63418]: INFO nova.compute.manager [req-d7a924cd-48c4-4aaa-b119-2d3a79c22e4e req-238bbb8b-c20e-403e-b61f-5ce66d61d95f service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Neutron deleted interface bf027be2-ee92-4c69-ad07-fe9393549d89; detaching it from the instance and deleting it from the info cache [ 862.006478] env[63418]: DEBUG nova.network.neutron [req-d7a924cd-48c4-4aaa-b119-2d3a79c22e4e req-238bbb8b-c20e-403e-b61f-5ce66d61d95f service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.167652] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.316280] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245092, 'name': ReconfigVM_Task, 'duration_secs': 0.300941} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.318840] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Reconfigured VM instance instance-0000004a to attach disk [datastore1] dcc94934-d8d2-46d9-85fd-e17c11a7a4d7/dcc94934-d8d2-46d9-85fd-e17c11a7a4d7.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.320093] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245093, 'name': CreateSnapshot_Task, 'duration_secs': 0.594332} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.323333] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32964e48-a2c1-4c4d-a362-59c349d06fc1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.325692] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 862.326621] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ba5fe4-1296-4683-b229-aa46cb39f463 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.348036] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 862.348036] env[63418]: value = "task-1245095" [ 862.348036] env[63418]: _type = "Task" [ 862.348036] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.358886] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245095, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.384947] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.476329] env[63418]: DEBUG nova.network.neutron [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.508548] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-115864cf-94c8-49b7-ad2a-4fbb8de0f53f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.519564] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657d40b2-6001-41d0-b43d-30667ed81ad2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.560036] env[63418]: DEBUG nova.compute.manager [req-d7a924cd-48c4-4aaa-b119-2d3a79c22e4e req-238bbb8b-c20e-403e-b61f-5ce66d61d95f service nova] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Detach interface failed, port_id=bf027be2-ee92-4c69-ad07-fe9393549d89, reason: Instance e2dcb8b8-e778-4202-8808-6a8535e2f1b4 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 862.566157] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9133e8b-8ba1-4eb2-ac47-4c500df6414b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.576355] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169102b1-418d-4bf3-bb23-8e85a25cf0e2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.615668] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc8964b-fb2c-4901-8cb1-74bd63c7ef84 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.625089] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f8a7c6-dba1-4529-9a23-28fcfea13d19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.640748] env[63418]: DEBUG nova.compute.provider_tree [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.860405] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 862.861105] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dee77d02-d6dd-49b0-b4c0-d3346effb2a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.880917] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245095, 'name': Rename_Task, 'duration_secs': 0.183205} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.881327] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 862.881327] env[63418]: value = "task-1245096" [ 862.881327] env[63418]: _type = "Task" [ 862.881327] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.882083] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.882431] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71198990-233e-4d5f-9551-ec9d0e8cd913 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.891610] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.897643] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 862.897643] env[63418]: value = "task-1245097" [ 862.897643] env[63418]: _type = "Task" [ 862.897643] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.901023] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245096, 'name': CloneVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.910151] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.985226] env[63418]: INFO nova.compute.manager [-] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Took 1.67 seconds to deallocate network for instance. [ 863.144337] env[63418]: DEBUG nova.scheduler.client.report [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 863.385325] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.394607] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245096, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.409854] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245097, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.492402] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.649984] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.954s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.652908] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.029s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.653363] env[63418]: DEBUG nova.objects.instance [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lazy-loading 'resources' on Instance uuid 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.715475] env[63418]: INFO nova.network.neutron [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 863.889445] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.899303] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245096, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.911027] env[63418]: DEBUG oslo_vmware.api [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245097, 'name': PowerOnVM_Task, 'duration_secs': 0.978815} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.911283] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.911518] env[63418]: INFO nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Took 8.34 seconds to spawn the instance on the hypervisor. [ 863.911824] env[63418]: DEBUG nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.912649] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd81308-2120-480f-9e56-94a2762e7171 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.389261] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.402546] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245096, 'name': CloneVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.403340] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecacf1a-2cb6-46ad-b364-f76bc4b02277 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.410897] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f342e1-91b9-491a-9ce8-6ca1982716fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.455987] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3380d789-173c-4461-9e43-2a005b475d21 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.462201] env[63418]: INFO nova.compute.manager [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Took 36.33 seconds to build instance. [ 864.471658] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5f75f3-0ef8-4c96-be08-5634c37c1c89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.491789] env[63418]: DEBUG nova.compute.provider_tree [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.890133] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.902124] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245096, 'name': CloneVM_Task, 'duration_secs': 1.606141} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.902124] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Created linked-clone VM from snapshot [ 864.902124] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de951cec-2c2e-46c1-82b4-a09a5b23396f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.909551] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Uploading image 07894cb3-0997-4b70-8729-287b7917ad7d {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 864.932151] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 864.932151] env[63418]: value = "vm-268471" [ 864.932151] env[63418]: _type = "VirtualMachine" [ 864.932151] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 864.932151] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8025d5ab-4adb-443a-971f-1d96f4e3fe66 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.939243] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease: (returnval){ [ 864.939243] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523a319f-a1e3-d21c-762a-c65f2ff56d85" [ 864.939243] env[63418]: _type = "HttpNfcLease" [ 864.939243] env[63418]: } obtained for exporting VM: (result){ [ 864.939243] env[63418]: value = "vm-268471" [ 864.939243] env[63418]: _type = "VirtualMachine" [ 864.939243] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 864.939544] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the lease: (returnval){ [ 864.939544] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523a319f-a1e3-d21c-762a-c65f2ff56d85" [ 864.939544] env[63418]: _type = "HttpNfcLease" [ 864.939544] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 864.948318] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 864.948318] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523a319f-a1e3-d21c-762a-c65f2ff56d85" [ 864.948318] env[63418]: _type = "HttpNfcLease" [ 864.948318] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 864.971014] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f1658d37-8ab4-4946-aae0-76a546151a3d tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.853s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.971395] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.971625] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.971866] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.972017] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.972220] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.974303] env[63418]: INFO nova.compute.manager [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Terminating instance [ 864.995499] env[63418]: DEBUG nova.scheduler.client.report [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 865.249221] env[63418]: DEBUG nova.compute.manager [req-122ef356-da80-4267-95f2-e39981ba617a req-f03fb41b-3ca2-4d59-966a-e301884d3804 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-vif-plugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 865.249830] env[63418]: DEBUG oslo_concurrency.lockutils [req-122ef356-da80-4267-95f2-e39981ba617a req-f03fb41b-3ca2-4d59-966a-e301884d3804 service nova] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.250269] env[63418]: DEBUG oslo_concurrency.lockutils [req-122ef356-da80-4267-95f2-e39981ba617a req-f03fb41b-3ca2-4d59-966a-e301884d3804 service nova] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.250568] env[63418]: DEBUG oslo_concurrency.lockutils [req-122ef356-da80-4267-95f2-e39981ba617a req-f03fb41b-3ca2-4d59-966a-e301884d3804 service nova] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.250984] env[63418]: DEBUG nova.compute.manager [req-122ef356-da80-4267-95f2-e39981ba617a req-f03fb41b-3ca2-4d59-966a-e301884d3804 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] No waiting events found dispatching network-vif-plugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 865.251237] env[63418]: WARNING nova.compute.manager [req-122ef356-da80-4267-95f2-e39981ba617a req-f03fb41b-3ca2-4d59-966a-e301884d3804 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received unexpected event network-vif-plugged-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 for instance with vm_state shelved_offloaded and task_state spawning. [ 865.387598] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.389786] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.390036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.390148] env[63418]: DEBUG nova.network.neutron [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.451188] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 865.451188] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523a319f-a1e3-d21c-762a-c65f2ff56d85" [ 865.451188] env[63418]: _type = "HttpNfcLease" [ 865.451188] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 865.451555] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 865.451555] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523a319f-a1e3-d21c-762a-c65f2ff56d85" [ 865.451555] env[63418]: _type = "HttpNfcLease" [ 865.451555] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 865.452771] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69d2c70-fb4d-44c9-88be-b71ed99fe683 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.464626] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263f199-1205-d72b-974b-07edf0216c45/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 865.464735] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263f199-1205-d72b-974b-07edf0216c45/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 865.532178] env[63418]: DEBUG nova.compute.manager [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 865.532533] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.533214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.536821] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e3762c-401d-4e4c-ada0-b2fc95cfeb25 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.539943] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.426s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.541855] env[63418]: INFO nova.compute.claims [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.552601] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.552902] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91330ab8-e1f3-47d1-b921-3ede97fc4a89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.559332] env[63418]: INFO nova.scheduler.client.report [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Deleted allocations for instance 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00 [ 865.562550] env[63418]: DEBUG oslo_vmware.api [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 865.562550] env[63418]: value = "task-1245099" [ 865.562550] env[63418]: _type = "Task" [ 865.562550] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.577643] env[63418]: DEBUG oslo_vmware.api [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.581300] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-030efd8e-64e6-46fb-942f-3876737c5e5f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.889152] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.076712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-40c64ce3-f305-4910-a8a1-eb25a3179b55 tempest-ServersTestManualDisk-8582898 tempest-ServersTestManualDisk-8582898-project-member] Lock "7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.127s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.091403] env[63418]: DEBUG oslo_vmware.api [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245099, 'name': PowerOffVM_Task, 'duration_secs': 0.209279} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.094139] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.095501] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.097223] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24c06a85-d17a-4508-bebb-b3c3f3c0793a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.175998] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.176362] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.176612] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleting the datastore file [datastore1] dcc94934-d8d2-46d9-85fd-e17c11a7a4d7 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.177389] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e21aef65-d2e9-46b3-91c4-13176ab86dd3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.186698] env[63418]: DEBUG oslo_vmware.api [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 866.186698] env[63418]: value = "task-1245101" [ 866.186698] env[63418]: _type = "Task" [ 866.186698] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.198689] env[63418]: DEBUG oslo_vmware.api [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.335029] env[63418]: DEBUG nova.network.neutron [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.394372] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.699148] env[63418]: DEBUG oslo_vmware.api [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256145} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.699519] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.699799] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.699992] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.700274] env[63418]: INFO nova.compute.manager [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Took 1.17 seconds to destroy the instance on the hypervisor. [ 866.700604] env[63418]: DEBUG oslo.service.loopingcall [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.700884] env[63418]: DEBUG nova.compute.manager [-] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 866.701123] env[63418]: DEBUG nova.network.neutron [-] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 866.790181] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a1f321-f84f-4d54-a214-be7ff6ba1283 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.799953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb321824-7652-48c8-ae27-5209b989f07a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.837756] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193a1c75-3cc8-4dd8-97b7-2e6a140bee26 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.841094] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.850674] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6e5352-dc58-41f9-83e9-748384f7a6f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.872499] env[63418]: DEBUG nova.compute.provider_tree [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.889509] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.378161] env[63418]: DEBUG nova.scheduler.client.report [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 867.397552] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.676772] env[63418]: DEBUG nova.network.neutron [-] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.890354] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.890923] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 867.897214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.813s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.898765] env[63418]: INFO nova.compute.claims [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.901450] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.183707] env[63418]: INFO nova.compute.manager [-] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Took 1.48 seconds to deallocate network for instance. [ 868.395323] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.404133] env[63418]: DEBUG nova.compute.utils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.405882] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.509s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.406706] env[63418]: DEBUG nova.compute.utils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Instance 394744c0-32eb-4c72-8295-5787706d32b4 could not be found. {{(pid=63418) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 868.408129] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 868.408338] env[63418]: DEBUG nova.network.neutron [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.410223] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.791s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.411755] env[63418]: INFO nova.compute.claims [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.414827] env[63418]: DEBUG nova.compute.manager [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Instance disappeared during build. {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2519}} [ 868.415034] env[63418]: DEBUG nova.compute.manager [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Unplugging VIFs for instance {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 868.415283] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Acquiring lock "refresh_cache-394744c0-32eb-4c72-8295-5787706d32b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.415466] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Acquired lock "refresh_cache-394744c0-32eb-4c72-8295-5787706d32b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.416581] env[63418]: DEBUG nova.network.neutron [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.465522] env[63418]: DEBUG nova.policy [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8609003874d4c41803c2567237e934e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b62f8bfe526440d1b882a38efa11795b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.691172] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.762860] env[63418]: DEBUG nova.network.neutron [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Successfully created port: 606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.898866] env[63418]: DEBUG oslo_vmware.api [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245094, 'name': ReconfigVM_Task, 'duration_secs': 6.800173} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.899225] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.899574] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Reconfigured VM to detach interface {{(pid=63418) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 868.909217] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 868.920485] env[63418]: DEBUG nova.compute.utils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Can not refresh info_cache because instance was not found {{(pid=63418) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 869.160438] env[63418]: DEBUG nova.network.neutron [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.246681] env[63418]: DEBUG nova.network.neutron [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.614847] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b908a0-9989-40e5-83d1-3ff24f33e1c2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.623690] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf25b5f5-125d-4ff6-bed2-47f76556d2bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.654618] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c02ee35-9c42-4a59-bcc8-424bec259994 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.663382] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5139bba-1a97-4d52-b905-f4e675072465 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.677736] env[63418]: DEBUG nova.compute.provider_tree [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.751297] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Releasing lock "refresh_cache-394744c0-32eb-4c72-8295-5787706d32b4" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.751601] env[63418]: DEBUG nova.compute.manager [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63418) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 869.751916] env[63418]: DEBUG nova.compute.manager [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] [instance: 394744c0-32eb-4c72-8295-5787706d32b4] Skipping network deallocation for instance since networking was not requested. {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2291}} [ 869.923243] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 870.181756] env[63418]: DEBUG nova.scheduler.client.report [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 870.191363] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='babc6b8aec70244cef7fe5452d54d767',container_format='bare',created_at=2024-10-10T13:42:38Z,direct_url=,disk_format='vmdk',id=14d64c0d-64ca-4b9e-9590-07ca59045ade,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2146101316-shelved',owner='0e9f537407b84d50a49600de59e72c86',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-10T13:42:56Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.191593] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.191665] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.191851] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.192027] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.195788] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.196060] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.196246] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.196446] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.196686] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.196915] env[63418]: DEBUG nova.virt.hardware [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.199372] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6e3d6e-b893-4ec4-a2f5-a6e38e35c246 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.210969] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.211240] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.211433] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.211640] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.211792] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.211939] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.212232] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.212409] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.212583] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.212831] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.212913] env[63418]: DEBUG nova.virt.hardware [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.214356] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3254a680-5f67-44ec-9310-9cd674a91d45 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.220753] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc946cdf-f6b5-4ebb-bbef-e6a58d041e27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.234742] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237e16e-5036-4b50-804a-34917498bb1c/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.246504] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c54dcc-44a4-445f-824e-4a03f84f2c3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.251676] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130e215c-a859-4cae-abfc-531e369a3a04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.255495] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:37:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0263ccbe-8541-4cf4-bd2a-0e9b517d6f29', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.263380] env[63418]: DEBUG oslo.service.loopingcall [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.264753] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.268019] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cfae518-e7cc-4373-863c-6555aba286ad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.293264] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237e16e-5036-4b50-804a-34917498bb1c/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 870.293440] env[63418]: ERROR oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237e16e-5036-4b50-804a-34917498bb1c/disk-0.vmdk due to incomplete transfer. [ 870.294382] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a7a2339d-d14e-41ae-890f-a169b8558f16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.300294] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.300294] env[63418]: value = "task-1245102" [ 870.300294] env[63418]: _type = "Task" [ 870.300294] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.305187] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237e16e-5036-4b50-804a-34917498bb1c/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 870.305403] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Uploaded image 218c0554-98fc-478a-bef5-e1a4088a8355 to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 870.307527] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 870.308172] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bc6931bd-9129-44e8-a845-c6d9452d66cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.312912] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245102, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.319539] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 870.319539] env[63418]: value = "task-1245103" [ 870.319539] env[63418]: _type = "Task" [ 870.319539] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.329164] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245103, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.412896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.413149] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.413358] env[63418]: DEBUG nova.network.neutron [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.427061] env[63418]: DEBUG nova.network.neutron [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Successfully updated port: 606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.703972] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.704626] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 870.707809] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.495s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.708074] env[63418]: DEBUG nova.objects.instance [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'resources' on Instance uuid 61af37a5-8c18-47a7-817a-eb3b332b3725 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.785388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6fc70ff4-e5f3-4500-95f7-f99dabc2864f tempest-ServersListShow296Test-728293733 tempest-ServersListShow296Test-728293733-project-member] Lock "394744c0-32eb-4c72-8295-5787706d32b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.723s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.815681] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245102, 'name': CreateVM_Task, 'duration_secs': 0.479261} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.815959] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.816750] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.817029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.817502] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.817832] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b3cc639-2643-4ebd-8204-48b8e0bed647 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.829529] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 870.829529] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527292e3-e3d0-58a8-2a26-6052c91d4299" [ 870.829529] env[63418]: _type = "Task" [ 870.829529] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.843747] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245103, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.854465] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.854465] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Processing image 14d64c0d-64ca-4b9e-9590-07ca59045ade {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.854465] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.854465] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.854465] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.854465] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5f31cfc-ce75-43dc-8176-a624e6e6222c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.868610] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.868838] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.869648] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a5170d2-5d81-481e-a15a-915d49cc0d4b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.876968] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 870.876968] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b32cad-f19d-5b05-ce11-2ef925bad557" [ 870.876968] env[63418]: _type = "Task" [ 870.876968] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.891112] env[63418]: DEBUG nova.compute.manager [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 870.891469] env[63418]: DEBUG nova.compute.manager [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing instance network info cache due to event network-changed-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 870.891823] env[63418]: DEBUG oslo_concurrency.lockutils [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.892013] env[63418]: DEBUG oslo_concurrency.lockutils [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.892253] env[63418]: DEBUG nova.network.neutron [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Refreshing network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.897473] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b32cad-f19d-5b05-ce11-2ef925bad557, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.930878] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "refresh_cache-384b4ae7-ad55-4d44-bd10-3b82534c9703" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.930878] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquired lock "refresh_cache-384b4ae7-ad55-4d44-bd10-3b82534c9703" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.930878] env[63418]: DEBUG nova.network.neutron [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.178637] env[63418]: DEBUG nova.compute.manager [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-vif-deleted-426ae826-8d40-4981-b7d4-280705c0ac16 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 871.178838] env[63418]: INFO nova.compute.manager [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Neutron deleted interface 426ae826-8d40-4981-b7d4-280705c0ac16; detaching it from the instance and deleting it from the info cache [ 871.179155] env[63418]: DEBUG nova.network.neutron [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "124ef09c-c406-439c-89c0-b00269da9af4", "address": "fa:16:3e:ed:1d:47", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124ef09c-c4", "ovs_interfaceid": "124ef09c-c406-439c-89c0-b00269da9af4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.205218] env[63418]: DEBUG nova.compute.manager [req-2dd01dff-c309-4afc-a438-aa6114a0709a req-c7c53675-1dce-430f-a858-ca8825788dc8 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Received event network-vif-plugged-606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 871.205218] env[63418]: DEBUG oslo_concurrency.lockutils [req-2dd01dff-c309-4afc-a438-aa6114a0709a req-c7c53675-1dce-430f-a858-ca8825788dc8 service nova] Acquiring lock "384b4ae7-ad55-4d44-bd10-3b82534c9703-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.205218] env[63418]: DEBUG oslo_concurrency.lockutils [req-2dd01dff-c309-4afc-a438-aa6114a0709a req-c7c53675-1dce-430f-a858-ca8825788dc8 service nova] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.205218] env[63418]: DEBUG oslo_concurrency.lockutils [req-2dd01dff-c309-4afc-a438-aa6114a0709a req-c7c53675-1dce-430f-a858-ca8825788dc8 service nova] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.205218] env[63418]: DEBUG nova.compute.manager [req-2dd01dff-c309-4afc-a438-aa6114a0709a req-c7c53675-1dce-430f-a858-ca8825788dc8 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] No waiting events found dispatching network-vif-plugged-606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 871.205218] env[63418]: WARNING nova.compute.manager [req-2dd01dff-c309-4afc-a438-aa6114a0709a req-c7c53675-1dce-430f-a858-ca8825788dc8 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Received unexpected event network-vif-plugged-606a5028-84ee-4542-aa24-d9a709985395 for instance with vm_state building and task_state spawning. [ 871.214283] env[63418]: DEBUG nova.compute.utils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.214283] env[63418]: DEBUG nova.objects.instance [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'numa_topology' on Instance uuid 61af37a5-8c18-47a7-817a-eb3b332b3725 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.216165] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 871.216500] env[63418]: DEBUG nova.network.neutron [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.293334] env[63418]: DEBUG nova.policy [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2dcdc53fd7cc4555a4eecb6bf7564b55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4759c1d6199344c49063d35e422d6324', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.330553] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245103, 'name': Destroy_Task, 'duration_secs': 0.581947} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.331465] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroyed the VM [ 871.331838] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 871.332225] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1dbf121e-9ca9-48c6-98f7-4513ce8d5ebb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.341737] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 871.341737] env[63418]: value = "task-1245104" [ 871.341737] env[63418]: _type = "Task" [ 871.341737] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.345491] env[63418]: INFO nova.network.neutron [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Port 426ae826-8d40-4981-b7d4-280705c0ac16 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 871.345867] env[63418]: INFO nova.network.neutron [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Port 124ef09c-c406-439c-89c0-b00269da9af4 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 871.346478] env[63418]: DEBUG nova.network.neutron [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [{"id": "1eca951d-2df6-4179-aa0c-6babf13838f3", "address": "fa:16:3e:1f:49:33", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eca951d-2d", "ovs_interfaceid": "1eca951d-2df6-4179-aa0c-6babf13838f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.355290] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245104, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.394637] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 871.395103] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Fetch image to [datastore1] OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2/OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 871.395423] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Downloading stream optimized image 14d64c0d-64ca-4b9e-9590-07ca59045ade to [datastore1] OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2/OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2.vmdk on the data store datastore1 as vApp {{(pid=63418) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 871.395757] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Downloading image file data 14d64c0d-64ca-4b9e-9590-07ca59045ade to the ESX as VM named 'OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2' {{(pid=63418) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 871.405494] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.405494] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.488613] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 871.488613] env[63418]: value = "resgroup-9" [ 871.488613] env[63418]: _type = "ResourcePool" [ 871.488613] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 871.488613] env[63418]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ea93eece-14bd-4b30-838e-9e77f046da4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.514477] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease: (returnval){ [ 871.514477] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a09ef8-06f0-c32b-0a55-3cdac308f976" [ 871.514477] env[63418]: _type = "HttpNfcLease" [ 871.514477] env[63418]: } obtained for vApp import into resource pool (val){ [ 871.514477] env[63418]: value = "resgroup-9" [ 871.514477] env[63418]: _type = "ResourcePool" [ 871.514477] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 871.515057] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the lease: (returnval){ [ 871.515057] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a09ef8-06f0-c32b-0a55-3cdac308f976" [ 871.515057] env[63418]: _type = "HttpNfcLease" [ 871.515057] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 871.519268] env[63418]: DEBUG nova.network.neutron [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.527269] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 871.527269] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a09ef8-06f0-c32b-0a55-3cdac308f976" [ 871.527269] env[63418]: _type = "HttpNfcLease" [ 871.527269] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 871.690420] env[63418]: DEBUG oslo_concurrency.lockutils [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.690606] env[63418]: DEBUG oslo_concurrency.lockutils [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] Acquired lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.691624] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e85a5db-7721-4b5c-baaa-900792c4fdd3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.712635] env[63418]: DEBUG oslo_concurrency.lockutils [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] Releasing lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.713148] env[63418]: WARNING nova.compute.manager [req-83c94957-ff10-4c09-a031-0dba71a20ecd req-96b52e48-a099-4ddf-805e-1b16c12fab39 service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Detach interface failed, port_id=426ae826-8d40-4981-b7d4-280705c0ac16, reason: No device with interface-id 426ae826-8d40-4981-b7d4-280705c0ac16 exists on VM: nova.exception.NotFound: No device with interface-id 426ae826-8d40-4981-b7d4-280705c0ac16 exists on VM [ 871.717081] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 871.721034] env[63418]: DEBUG nova.objects.base [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Object Instance<61af37a5-8c18-47a7-817a-eb3b332b3725> lazy-loaded attributes: resources,numa_topology {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 871.736707] env[63418]: DEBUG nova.network.neutron [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updated VIF entry in instance network info cache for port 0263ccbe-8541-4cf4-bd2a-0e9b517d6f29. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.737389] env[63418]: DEBUG nova.network.neutron [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [{"id": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "address": "fa:16:3e:7b:37:75", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0263ccbe-85", "ovs_interfaceid": "0263ccbe-8541-4cf4-bd2a-0e9b517d6f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.783060] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.783351] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.783572] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.785467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.785467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.790668] env[63418]: INFO nova.compute.manager [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Terminating instance [ 871.820287] env[63418]: DEBUG nova.network.neutron [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Updating instance_info_cache with network_info: [{"id": "606a5028-84ee-4542-aa24-d9a709985395", "address": "fa:16:3e:6b:d8:ce", "network": {"id": "5e0d92d4-7a7a-46cf-ab33-6130b90a7c84", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-921599569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b62f8bfe526440d1b882a38efa11795b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "462a7219-4deb-4225-9cf7-3131ef280363", "external-id": "nsx-vlan-transportzone-468", "segmentation_id": 468, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606a5028-84", "ovs_interfaceid": "606a5028-84ee-4542-aa24-d9a709985395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.850447] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.854345] env[63418]: DEBUG nova.network.neutron [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Successfully created port: ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.860671] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245104, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.907523] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 872.025229] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 872.025229] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a09ef8-06f0-c32b-0a55-3cdac308f976" [ 872.025229] env[63418]: _type = "HttpNfcLease" [ 872.025229] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 872.028931] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6fc970-41f4-459f-9ff4-2554055ed9dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.041506] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0489b38-5968-41a2-9c4e-e2bdf10d3f07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.074468] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ba5296-8e17-4ffb-bb0d-9d142b595c50 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.083287] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c46a9d-92fb-44fb-9fdd-8797132798b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.100124] env[63418]: DEBUG nova.compute.provider_tree [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.239835] env[63418]: DEBUG oslo_concurrency.lockutils [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.240196] env[63418]: DEBUG nova.compute.manager [req-8a7aca64-804e-445d-a1ad-d6e4b977ebf9 req-b6574f9a-b353-40b4-a63c-e3311ae04632 service nova] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Received event network-vif-deleted-c3c91249-1c73-4686-baaa-8a7fe085b98d {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 872.305763] env[63418]: DEBUG nova.compute.manager [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 872.306105] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.307115] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a762dc-84de-4c53-a86d-eefbf50423cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.316072] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.316455] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa00889e-a349-42c0-a769-c00414ee25c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.323865] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Releasing lock "refresh_cache-384b4ae7-ad55-4d44-bd10-3b82534c9703" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.324240] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Instance network_info: |[{"id": "606a5028-84ee-4542-aa24-d9a709985395", "address": "fa:16:3e:6b:d8:ce", "network": {"id": "5e0d92d4-7a7a-46cf-ab33-6130b90a7c84", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-921599569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b62f8bfe526440d1b882a38efa11795b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "462a7219-4deb-4225-9cf7-3131ef280363", "external-id": "nsx-vlan-transportzone-468", "segmentation_id": 468, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606a5028-84", "ovs_interfaceid": "606a5028-84ee-4542-aa24-d9a709985395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 872.324599] env[63418]: DEBUG oslo_vmware.api [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 872.324599] env[63418]: value = "task-1245106" [ 872.324599] env[63418]: _type = "Task" [ 872.324599] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.324929] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:d8:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '462a7219-4deb-4225-9cf7-3131ef280363', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '606a5028-84ee-4542-aa24-d9a709985395', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.332905] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Creating folder: Project (b62f8bfe526440d1b882a38efa11795b). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.333697] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-925f1b6a-64d6-420d-a6a8-9b310b0292aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.345313] env[63418]: DEBUG oslo_vmware.api [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245106, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.350162] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Created folder: Project (b62f8bfe526440d1b882a38efa11795b) in parent group-v268354. [ 872.350162] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Creating folder: Instances. Parent ref: group-v268474. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.350729] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d88d1383-82b7-49f2-b841-83f3c21b84e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.356108] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6a99d1c6-064f-44fe-912f-91fc28b36291 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-eaa13276-1fb1-47e2-ad1f-445bc9f4c98f-426ae826-8d40-4981-b7d4-280705c0ac16" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.103s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.357181] env[63418]: DEBUG oslo_vmware.api [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245104, 'name': RemoveSnapshot_Task, 'duration_secs': 0.550597} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.357650] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 872.357875] env[63418]: INFO nova.compute.manager [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 14.39 seconds to snapshot the instance on the hypervisor. [ 872.361252] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Created folder: Instances in parent group-v268474. [ 872.361493] env[63418]: DEBUG oslo.service.loopingcall [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.363787] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.363787] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da93a62e-acfa-4940-a3f5-a5fb5bc16480 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.384945] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.384945] env[63418]: value = "task-1245109" [ 872.384945] env[63418]: _type = "Task" [ 872.384945] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.392933] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245109, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.433264] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.525283] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 872.525283] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a09ef8-06f0-c32b-0a55-3cdac308f976" [ 872.525283] env[63418]: _type = "HttpNfcLease" [ 872.525283] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 872.525569] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 872.525569] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a09ef8-06f0-c32b-0a55-3cdac308f976" [ 872.525569] env[63418]: _type = "HttpNfcLease" [ 872.525569] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 872.526662] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7d4682-7d56-48e8-bf2f-cdc07da7c9d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.534985] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526badfd-7402-168b-1284-baf70014c8a2/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 872.535219] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526badfd-7402-168b-1284-baf70014c8a2/disk-0.vmdk. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 872.601710] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e8c7f863-dc50-46da-a699-b56d2c3649d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.605497] env[63418]: DEBUG nova.scheduler.client.report [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 872.635443] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263f199-1205-d72b-974b-07edf0216c45/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 872.637767] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a81930e-d5ee-43ac-9a81-c11a8bbc99cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.644806] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263f199-1205-d72b-974b-07edf0216c45/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 872.645022] env[63418]: ERROR oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263f199-1205-d72b-974b-07edf0216c45/disk-0.vmdk due to incomplete transfer. [ 872.645271] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d03eb46f-2d1c-4e4b-b62b-8a6dc43a6144 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.664362] env[63418]: DEBUG oslo_vmware.rw_handles [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263f199-1205-d72b-974b-07edf0216c45/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 872.664726] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Uploaded image 07894cb3-0997-4b70-8729-287b7917ad7d to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 872.666812] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 872.667136] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e376dc5a-b27c-49b9-83d2-a8f441670644 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.690113] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 872.690113] env[63418]: value = "task-1245110" [ 872.690113] env[63418]: _type = "Task" [ 872.690113] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.699125] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245110, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.731226] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 872.755397] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.755723] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.755896] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.756097] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.756256] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.756450] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.756703] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.756910] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.757117] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.757329] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.757474] env[63418]: DEBUG nova.virt.hardware [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.758488] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbd9792-66ac-4749-8d17-e1a157e5132b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.771457] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8aeb0e-1985-4457-9505-d66f482da7a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.843458] env[63418]: DEBUG oslo_vmware.api [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245106, 'name': PowerOffVM_Task, 'duration_secs': 0.213446} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.843766] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.843972] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.844268] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84b9c8cb-89e5-44e7-91ef-21ae9b596635 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.898113] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245109, 'name': CreateVM_Task, 'duration_secs': 0.381658} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.903922] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.905054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.905301] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.905894] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.907043] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68e040ac-b90a-4031-a28a-1f70bf4ee27a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.911059] env[63418]: DEBUG nova.compute.manager [None req-b7aee40a-bbef-4e57-8ae2-5575e8fb6cec tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Found 1 images (rotation: 2) {{(pid=63418) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 872.917598] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 872.917598] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520d27c3-67b3-2bf3-824b-05fa6a1a78a6" [ 872.917598] env[63418]: _type = "Task" [ 872.917598] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.929539] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520d27c3-67b3-2bf3-824b-05fa6a1a78a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.967615] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.967915] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.968124] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleting the datastore file [datastore2] eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.968644] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04189d7b-39a2-493c-a3f1-17b53b006a2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.978261] env[63418]: DEBUG oslo_vmware.api [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 872.978261] env[63418]: value = "task-1245112" [ 872.978261] env[63418]: _type = "Task" [ 872.978261] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.993685] env[63418]: DEBUG oslo_vmware.api [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.112330] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.404s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.117667] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.391s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.117802] env[63418]: DEBUG nova.objects.instance [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63418) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 873.212027] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245110, 'name': Destroy_Task, 'duration_secs': 0.36471} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.216318] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Destroyed the VM [ 873.216318] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 873.216318] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-eaa0a5d2-6f18-45ba-b04e-acc76506112d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.231615] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 873.231615] env[63418]: value = "task-1245113" [ 873.231615] env[63418]: _type = "Task" [ 873.231615] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.236902] env[63418]: DEBUG nova.compute.manager [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Received event network-changed-606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 873.237128] env[63418]: DEBUG nova.compute.manager [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Refreshing instance network info cache due to event network-changed-606a5028-84ee-4542-aa24-d9a709985395. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 873.237358] env[63418]: DEBUG oslo_concurrency.lockutils [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] Acquiring lock "refresh_cache-384b4ae7-ad55-4d44-bd10-3b82534c9703" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.237529] env[63418]: DEBUG oslo_concurrency.lockutils [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] Acquired lock "refresh_cache-384b4ae7-ad55-4d44-bd10-3b82534c9703" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.237815] env[63418]: DEBUG nova.network.neutron [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Refreshing network info cache for port 606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.257024] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245113, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.431304] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520d27c3-67b3-2bf3-824b-05fa6a1a78a6, 'name': SearchDatastore_Task, 'duration_secs': 0.011105} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.433862] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.434237] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.434601] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.434824] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.435035] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.435465] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-129cc052-30f5-4753-b47d-9616abba0f7a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.450285] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.450285] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.450285] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8232e2bd-6b66-4bfa-ba62-c9abd4b0ecc2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.458600] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 873.458600] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e64dd4-17b9-285a-090c-a48063e46772" [ 873.458600] env[63418]: _type = "Task" [ 873.458600] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.468514] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e64dd4-17b9-285a-090c-a48063e46772, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.494338] env[63418]: DEBUG oslo_vmware.api [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204108} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.497583] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.497854] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.498064] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.498323] env[63418]: INFO nova.compute.manager [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Took 1.19 seconds to destroy the instance on the hypervisor. [ 873.501780] env[63418]: DEBUG oslo.service.loopingcall [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.501780] env[63418]: DEBUG nova.compute.manager [-] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 873.501780] env[63418]: DEBUG nova.network.neutron [-] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.561089] env[63418]: DEBUG nova.network.neutron [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Successfully updated port: ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.569897] env[63418]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 124ef09c-c406-439c-89c0-b00269da9af4 could not be found.", "detail": ""}} {{(pid=63418) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 873.570199] env[63418]: DEBUG nova.network.neutron [-] Unable to show port 124ef09c-c406-439c-89c0-b00269da9af4 as it no longer exists. {{(pid=63418) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 873.631603] env[63418]: DEBUG oslo_concurrency.lockutils [None req-02302a52-a840-4091-b8d4-97d2ac2506dc tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.489s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.632747] env[63418]: DEBUG oslo_concurrency.lockutils [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 15.693s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.633033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.633288] env[63418]: DEBUG oslo_concurrency.lockutils [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.633499] env[63418]: DEBUG oslo_concurrency.lockutils [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.638226] env[63418]: INFO nova.compute.manager [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Terminating instance [ 873.680442] env[63418]: DEBUG nova.compute.manager [req-d3e5e6b8-4a3b-4c67-bb3a-9498be9a4a80 req-7dd53227-0111-44a1-975f-95bfaf2c5aba service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-vif-deleted-124ef09c-c406-439c-89c0-b00269da9af4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 873.752851] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "bf091892-fefb-49dd-9416-708a06b35798" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.753118] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "bf091892-fefb-49dd-9416-708a06b35798" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.757678] env[63418]: DEBUG oslo_vmware.api [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245113, 'name': RemoveSnapshot_Task, 'duration_secs': 0.496573} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.757678] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 873.757678] env[63418]: INFO nova.compute.manager [None req-97823767-7d43-4126-97e1-ff9be93c5ca0 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Took 12.61 seconds to snapshot the instance on the hypervisor. [ 873.885828] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 873.885828] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526badfd-7402-168b-1284-baf70014c8a2/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 873.886737] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2716726b-a410-4414-9a4e-d65a3ffeb773 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.895130] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526badfd-7402-168b-1284-baf70014c8a2/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 873.895438] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526badfd-7402-168b-1284-baf70014c8a2/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 873.895731] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-029c75da-75d9-42c7-97d1-90827a294fee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.972124] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e64dd4-17b9-285a-090c-a48063e46772, 'name': SearchDatastore_Task, 'duration_secs': 0.011254} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.972838] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f634f94-4fb1-4e6c-b339-36f6d1333633 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.981859] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 873.981859] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c8efdc-70e5-c5c7-0c94-ab9055ecffda" [ 873.981859] env[63418]: _type = "Task" [ 873.981859] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.990919] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c8efdc-70e5-c5c7-0c94-ab9055ecffda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.065024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "refresh_cache-5ab04ddd-4cd6-4330-a9d7-081290c27586" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.065024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "refresh_cache-5ab04ddd-4cd6-4330-a9d7-081290c27586" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.065024] env[63418]: DEBUG nova.network.neutron [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.123409] env[63418]: DEBUG nova.network.neutron [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Updated VIF entry in instance network info cache for port 606a5028-84ee-4542-aa24-d9a709985395. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.123827] env[63418]: DEBUG nova.network.neutron [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Updating instance_info_cache with network_info: [{"id": "606a5028-84ee-4542-aa24-d9a709985395", "address": "fa:16:3e:6b:d8:ce", "network": {"id": "5e0d92d4-7a7a-46cf-ab33-6130b90a7c84", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-921599569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b62f8bfe526440d1b882a38efa11795b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "462a7219-4deb-4225-9cf7-3131ef280363", "external-id": "nsx-vlan-transportzone-468", "segmentation_id": 468, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606a5028-84", "ovs_interfaceid": "606a5028-84ee-4542-aa24-d9a709985395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.136870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b7281253-0c28-4954-9efd-f94b6a391626 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.138982] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.222s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.142842] env[63418]: DEBUG nova.compute.manager [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 874.142842] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.143258] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-72c97a45-a88b-4cb6-b82d-0fafad9349b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.147084] env[63418]: DEBUG nova.compute.manager [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 874.148081] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076a317e-a6f3-46dd-ac20-2feddb1c2183 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.156675] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b65cfb5-67cf-468c-8ec1-121732e03712 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.176022] env[63418]: DEBUG oslo_vmware.rw_handles [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526badfd-7402-168b-1284-baf70014c8a2/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 874.176071] env[63418]: INFO nova.virt.vmwareapi.images [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Downloaded image file data 14d64c0d-64ca-4b9e-9590-07ca59045ade [ 874.176867] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335fa32b-13fd-43fb-a7b0-c75fe3614073 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.188341] env[63418]: WARNING nova.virt.vmwareapi.vmops [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61af37a5-8c18-47a7-817a-eb3b332b3725 could not be found. [ 874.188567] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.188754] env[63418]: INFO nova.compute.manager [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Took 0.05 seconds to destroy the instance on the hypervisor. [ 874.189013] env[63418]: DEBUG oslo.service.loopingcall [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.199424] env[63418]: DEBUG nova.compute.manager [-] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 874.199535] env[63418]: DEBUG nova.network.neutron [-] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.203208] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2075e919-6851-4efb-8890-90804b288e85 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.229717] env[63418]: INFO nova.virt.vmwareapi.images [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] The imported VM was unregistered [ 874.232145] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 874.232450] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating directory with path [datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.232871] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13fbc37f-dd7f-46e7-bf97-c2eabc500b55 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.245107] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created directory with path [datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.245322] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2/OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2.vmdk to [datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk. {{(pid=63418) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 874.245589] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-160fa578-d107-4125-8a27-582b81a183fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.253222] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 874.253222] env[63418]: value = "task-1245115" [ 874.253222] env[63418]: _type = "Task" [ 874.253222] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.261433] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 874.269019] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.460558] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.460850] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.461099] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.461291] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.461462] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.469251] env[63418]: INFO nova.compute.manager [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Terminating instance [ 874.495783] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c8efdc-70e5-c5c7-0c94-ab9055ecffda, 'name': SearchDatastore_Task, 'duration_secs': 0.028616} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.496365] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.496682] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 384b4ae7-ad55-4d44-bd10-3b82534c9703/384b4ae7-ad55-4d44-bd10-3b82534c9703.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.496973] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11c04126-73a4-4028-ac5e-28d99a8ea85c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.507159] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 874.507159] env[63418]: value = "task-1245116" [ 874.507159] env[63418]: _type = "Task" [ 874.507159] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.521566] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.618038] env[63418]: DEBUG nova.network.neutron [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.626601] env[63418]: DEBUG oslo_concurrency.lockutils [req-ae05b2dc-57cf-4b55-845f-d98e9556ed86 req-e3dc1dd7-df17-48a0-a988-2dfd8d8c5779 service nova] Releasing lock "refresh_cache-384b4ae7-ad55-4d44-bd10-3b82534c9703" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.672061] env[63418]: INFO nova.compute.manager [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] instance snapshotting [ 874.672813] env[63418]: DEBUG nova.objects.instance [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'flavor' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.706351] env[63418]: DEBUG nova.network.neutron [-] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.771116] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.780942] env[63418]: DEBUG nova.network.neutron [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Updating instance_info_cache with network_info: [{"id": "ed4100a3-20c0-49ae-8fdd-778787b060aa", "address": "fa:16:3e:9c:5a:7c", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped4100a3-20", "ovs_interfaceid": "ed4100a3-20c0-49ae-8fdd-778787b060aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.784062] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.972471] env[63418]: DEBUG nova.network.neutron [-] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.973836] env[63418]: DEBUG nova.compute.manager [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 874.973976] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.975545] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdc7768-7e4b-456c-9cd5-87d257034f56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.989097] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.989446] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2c92f79-6c4f-482b-be56-8a6ad0ae8bcf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.017976] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.040847] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.041113] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.091939] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.092262] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.092471] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleting the datastore file [datastore2] 1f213c5b-de27-48d6-a1f8-fdf9b77359fd {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.092804] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94d13401-8383-40f5-98dd-8c3768f44cb5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.106541] env[63418]: DEBUG oslo_vmware.api [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 875.106541] env[63418]: value = "task-1245118" [ 875.106541] env[63418]: _type = "Task" [ 875.106541] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.121634] env[63418]: DEBUG oslo_vmware.api [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245118, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.167678] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.167852] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance eaa13276-1fb1-47e2-ad1f-445bc9f4c98f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.167947] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7b0c70aa-e2bc-4131-97b4-4e53a378940a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.168079] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.168256] env[63418]: WARNING nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance e2dcb8b8-e778-4202-8808-6a8535e2f1b4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 875.168383] env[63418]: WARNING nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance e19a60cd-4ec7-48ad-9042-f19b75353364 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 875.168578] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 1f213c5b-de27-48d6-a1f8-fdf9b77359fd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.168674] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance d76a008c-9bd9-420b-873d-4f7d7f25b8ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.168738] env[63418]: WARNING nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance dcc94934-d8d2-46d9-85fd-e17c11a7a4d7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 875.168863] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 384b4ae7-ad55-4d44-bd10-3b82534c9703 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.168985] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 5ab04ddd-4cd6-4330-a9d7-081290c27586 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.179023] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657ff38c-096a-47a9-947e-4f3824c3c312 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.206952] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f971df8e-257b-49ff-a30e-4682352133a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.209856] env[63418]: INFO nova.compute.manager [-] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Took 1.71 seconds to deallocate network for instance. [ 875.269812] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.283995] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "refresh_cache-5ab04ddd-4cd6-4330-a9d7-081290c27586" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.284433] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Instance network_info: |[{"id": "ed4100a3-20c0-49ae-8fdd-778787b060aa", "address": "fa:16:3e:9c:5a:7c", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped4100a3-20", "ovs_interfaceid": "ed4100a3-20c0-49ae-8fdd-778787b060aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 875.284876] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:5a:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed4100a3-20c0-49ae-8fdd-778787b060aa', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.292548] env[63418]: DEBUG oslo.service.loopingcall [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.292789] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.293034] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-938d7c1d-2775-4610-868c-c63b3a2e3d07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.320149] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.320149] env[63418]: value = "task-1245119" [ 875.320149] env[63418]: _type = "Task" [ 875.320149] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.333923] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245119, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.479907] env[63418]: INFO nova.compute.manager [-] [instance: 61af37a5-8c18-47a7-817a-eb3b332b3725] Took 1.28 seconds to deallocate network for instance. [ 875.522879] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.544033] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 875.622713] env[63418]: DEBUG oslo_vmware.api [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245118, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210886} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.623043] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.623236] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.623514] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.623767] env[63418]: INFO nova.compute.manager [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Took 0.65 seconds to destroy the instance on the hypervisor. [ 875.624154] env[63418]: DEBUG oslo.service.loopingcall [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.624448] env[63418]: DEBUG nova.compute.manager [-] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 875.624574] env[63418]: DEBUG nova.network.neutron [-] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.672726] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 586f51a0-90de-4f44-ac3e-758ceda9e316 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 875.715874] env[63418]: DEBUG nova.compute.manager [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Received event network-vif-plugged-ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 875.716308] env[63418]: DEBUG oslo_concurrency.lockutils [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] Acquiring lock "5ab04ddd-4cd6-4330-a9d7-081290c27586-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.716529] env[63418]: DEBUG oslo_concurrency.lockutils [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.716735] env[63418]: DEBUG oslo_concurrency.lockutils [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.716896] env[63418]: DEBUG nova.compute.manager [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] No waiting events found dispatching network-vif-plugged-ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 875.717083] env[63418]: WARNING nova.compute.manager [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Received unexpected event network-vif-plugged-ed4100a3-20c0-49ae-8fdd-778787b060aa for instance with vm_state building and task_state spawning. [ 875.717253] env[63418]: DEBUG nova.compute.manager [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Received event network-changed-ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 875.717419] env[63418]: DEBUG nova.compute.manager [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Refreshing instance network info cache due to event network-changed-ed4100a3-20c0-49ae-8fdd-778787b060aa. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 875.717712] env[63418]: DEBUG oslo_concurrency.lockutils [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] Acquiring lock "refresh_cache-5ab04ddd-4cd6-4330-a9d7-081290c27586" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.717910] env[63418]: DEBUG oslo_concurrency.lockutils [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] Acquired lock "refresh_cache-5ab04ddd-4cd6-4330-a9d7-081290c27586" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.718101] env[63418]: DEBUG nova.network.neutron [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Refreshing network info cache for port ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.721635] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.725587] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 875.726455] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-89a87eb2-6244-46ef-a3ea-30448e4b1df0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.740806] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 875.740806] env[63418]: value = "task-1245120" [ 875.740806] env[63418]: _type = "Task" [ 875.740806] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.755719] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245120, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.770393] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.835020] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245119, 'name': CreateVM_Task, 'duration_secs': 0.427333} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.835463] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.836194] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.836491] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.836756] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.837080] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-103163ea-ac3b-4d46-a557-44f02f51103d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.847438] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 875.847438] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528c8ab0-a61e-ac9d-0b4a-0a09ef3be3f9" [ 875.847438] env[63418]: _type = "Task" [ 875.847438] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.861571] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528c8ab0-a61e-ac9d-0b4a-0a09ef3be3f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.023309] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245116, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.071867] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.175970] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 876.258426] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245120, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.270236] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.362928] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528c8ab0-a61e-ac9d-0b4a-0a09ef3be3f9, 'name': SearchDatastore_Task, 'duration_secs': 0.090964} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.364039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.364039] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.364039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.364039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.364326] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.364450] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e5c7c62-1b14-4449-b8a3-ad84a8d3e4e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.385538] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.385742] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.386551] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77781234-1ee6-40dc-b55a-06adf4f65416 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.396515] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 876.396515] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5222ab31-524c-4c77-a0c5-846f8ebd29b7" [ 876.396515] env[63418]: _type = "Task" [ 876.396515] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.409828] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5222ab31-524c-4c77-a0c5-846f8ebd29b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.416225] env[63418]: DEBUG nova.network.neutron [-] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.452048] env[63418]: DEBUG nova.network.neutron [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Updated VIF entry in instance network info cache for port ed4100a3-20c0-49ae-8fdd-778787b060aa. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.452545] env[63418]: DEBUG nova.network.neutron [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Updating instance_info_cache with network_info: [{"id": "ed4100a3-20c0-49ae-8fdd-778787b060aa", "address": "fa:16:3e:9c:5a:7c", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped4100a3-20", "ovs_interfaceid": "ed4100a3-20c0-49ae-8fdd-778787b060aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.516437] env[63418]: DEBUG oslo_concurrency.lockutils [None req-47657c51-7dd1-4f15-ae53-3538a6e7c6ff tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "61af37a5-8c18-47a7-817a-eb3b332b3725" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.884s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.527395] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245116, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.679498] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 876.755365] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245120, 'name': CreateSnapshot_Task, 'duration_secs': 0.774939} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.755738] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 876.756685] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063cd97b-4e6f-4be8-909b-e561364d53ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.775628] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.907663] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5222ab31-524c-4c77-a0c5-846f8ebd29b7, 'name': SearchDatastore_Task, 'duration_secs': 0.093463} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.908553] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f734cc7-9563-479b-a944-5833f9aa2575 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.914340] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 876.914340] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52fbb5d3-c050-f2db-cc4c-1a5c56709069" [ 876.914340] env[63418]: _type = "Task" [ 876.914340] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.925519] env[63418]: INFO nova.compute.manager [-] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Took 1.30 seconds to deallocate network for instance. [ 876.925649] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52fbb5d3-c050-f2db-cc4c-1a5c56709069, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.955717] env[63418]: DEBUG oslo_concurrency.lockutils [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] Releasing lock "refresh_cache-5ab04ddd-4cd6-4330-a9d7-081290c27586" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.955717] env[63418]: DEBUG nova.compute.manager [req-622d18d0-9eef-44db-b196-8b38d5f795ce req-f85dafcf-6317-46a1-ae29-efa4dab5d57b service nova] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Received event network-vif-deleted-1eca951d-2df6-4179-aa0c-6babf13838f3 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 877.022011] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245116, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.394127} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.022343] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 384b4ae7-ad55-4d44-bd10-3b82534c9703/384b4ae7-ad55-4d44-bd10-3b82534c9703.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.022565] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.022820] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72453e95-6b56-4d62-b7e7-040044e46728 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.029792] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 877.029792] env[63418]: value = "task-1245121" [ 877.029792] env[63418]: _type = "Task" [ 877.029792] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.037675] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.183493] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance bf091892-fefb-49dd-9416-708a06b35798 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 877.270600] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245115, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.666927} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.271129] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2/OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2.vmdk to [datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk. [ 877.271642] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Cleaning up location [datastore1] OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 877.271840] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ae1d4823-42ef-46e3-a349-f539aaf9f2f2 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.272129] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71129295-0ece-4804-acbd-47a52f1b2eaa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.283211] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 877.284722] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7a14d452-efc4-43ab-86d9-2826a25cb832 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.287685] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 877.287685] env[63418]: value = "task-1245122" [ 877.287685] env[63418]: _type = "Task" [ 877.287685] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.295537] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 877.295537] env[63418]: value = "task-1245123" [ 877.295537] env[63418]: _type = "Task" [ 877.295537] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.300027] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.305829] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245123, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.425274] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52fbb5d3-c050-f2db-cc4c-1a5c56709069, 'name': SearchDatastore_Task, 'duration_secs': 0.015499} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.425597] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.425893] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5ab04ddd-4cd6-4330-a9d7-081290c27586/5ab04ddd-4cd6-4330-a9d7-081290c27586.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.426188] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2f0c393-a92d-4c28-bf26-223c1141f88e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.431309] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.433049] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 877.433049] env[63418]: value = "task-1245124" [ 877.433049] env[63418]: _type = "Task" [ 877.433049] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.441809] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.540182] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06281} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.540424] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.541254] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04bfc6a-d055-4e70-97cb-4bdb00746ee0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.564453] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 384b4ae7-ad55-4d44-bd10-3b82534c9703/384b4ae7-ad55-4d44-bd10-3b82534c9703.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.564793] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-889deb3a-70be-4036-9a93-1b6ce477accd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.585955] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 877.585955] env[63418]: value = "task-1245125" [ 877.585955] env[63418]: _type = "Task" [ 877.585955] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.599741] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245125, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.606343] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "054bdcde-55a6-4f86-a230-92b2115e47b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.606773] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.689038] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance dcb52a03-8f82-4b33-8dc6-3924140db8d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 877.689038] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 877.689328] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 877.743622] env[63418]: DEBUG nova.compute.manager [req-70ecbcf8-044e-47f8-a17c-fd4de06f2aea req-0475b669-d372-4661-b9d7-c8873eafec56 service nova] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Received event network-vif-deleted-4a552c2b-4e02-4200-b995-63e7101cf588 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 877.805649] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042772} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.806100] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.806343] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.806664] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk to [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.807546] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e1c187e-4ed3-429e-b8c3-a2d306e65585 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.814639] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245123, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.824241] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 877.824241] env[63418]: value = "task-1245126" [ 877.824241] env[63418]: _type = "Task" [ 877.824241] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.837219] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245126, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.949249] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245124, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.961569] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a834dbac-9394-40ec-bc98-8ee075d7581b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.970092] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc36b1c4-5dcd-4d9f-80bc-add25ee7cf57 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.005204] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b39cbf-43ab-43a2-af66-f46c47e41fb8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.014218] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ae165e-3b3b-4f97-8aed-5306ca270c01 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.028401] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.098829] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245125, 'name': ReconfigVM_Task, 'duration_secs': 0.489643} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.099117] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 384b4ae7-ad55-4d44-bd10-3b82534c9703/384b4ae7-ad55-4d44-bd10-3b82534c9703.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.099782] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8acdc217-95bf-49d0-84cc-224ab82967aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.109696] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 878.113988] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 878.113988] env[63418]: value = "task-1245127" [ 878.113988] env[63418]: _type = "Task" [ 878.113988] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.127241] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245127, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.315895] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245123, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.338738] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245126, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.453114] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245124, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544195} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.453114] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5ab04ddd-4cd6-4330-a9d7-081290c27586/5ab04ddd-4cd6-4330-a9d7-081290c27586.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.453751] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.454453] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-489080cd-6972-483e-8c1b-a22e0969c8f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.472430] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 878.472430] env[63418]: value = "task-1245128" [ 878.472430] env[63418]: _type = "Task" [ 878.472430] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.489441] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245128, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.532605] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 878.639305] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245127, 'name': Rename_Task, 'duration_secs': 0.224436} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.639305] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.639305] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-349e7303-cb62-4391-bd01-85bf4d676aac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.643247] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.651561] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 878.651561] env[63418]: value = "task-1245129" [ 878.651561] env[63418]: _type = "Task" [ 878.651561] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.665595] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.812387] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245123, 'name': CloneVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.839942] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245126, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.985085] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245128, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.30854} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.985408] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.986284] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022486ea-21fe-477c-90f2-a1dae6461e08 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.015355] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 5ab04ddd-4cd6-4330-a9d7-081290c27586/5ab04ddd-4cd6-4330-a9d7-081290c27586.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.015701] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91acc8d6-2daf-44d4-a6c2-5536ca1df224 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.038284] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 879.038592] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.900s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.039522] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.560s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.041146] env[63418]: INFO nova.compute.claims [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.045299] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 879.045299] env[63418]: value = "task-1245130" [ 879.045299] env[63418]: _type = "Task" [ 879.045299] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.060581] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245130, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.170064] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245129, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.313215] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245123, 'name': CloneVM_Task, 'duration_secs': 1.528185} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.313677] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created linked-clone VM from snapshot [ 879.315640] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ac9302-2955-4d09-9d78-9b31d8bc4d92 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.328143] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Uploading image 8d2a987b-43f8-48c5-a6e9-e586f49f246f {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 879.344701] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245126, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.355230] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 879.355230] env[63418]: value = "vm-268479" [ 879.355230] env[63418]: _type = "VirtualMachine" [ 879.355230] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 879.355540] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-98371c5b-4f8d-4d6d-a21a-7e595ecbf53b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.368621] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease: (returnval){ [ 879.368621] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0131e-8a6f-242a-9908-b452f6de6e9c" [ 879.368621] env[63418]: _type = "HttpNfcLease" [ 879.368621] env[63418]: } obtained for exporting VM: (result){ [ 879.368621] env[63418]: value = "vm-268479" [ 879.368621] env[63418]: _type = "VirtualMachine" [ 879.368621] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 879.368947] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the lease: (returnval){ [ 879.368947] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0131e-8a6f-242a-9908-b452f6de6e9c" [ 879.368947] env[63418]: _type = "HttpNfcLease" [ 879.368947] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 879.380443] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 879.380443] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0131e-8a6f-242a-9908-b452f6de6e9c" [ 879.380443] env[63418]: _type = "HttpNfcLease" [ 879.380443] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 879.563509] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245130, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.668929] env[63418]: DEBUG oslo_vmware.api [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245129, 'name': PowerOnVM_Task, 'duration_secs': 0.575814} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.669234] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.669514] env[63418]: INFO nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Took 9.75 seconds to spawn the instance on the hypervisor. [ 879.669702] env[63418]: DEBUG nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.670633] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d61ee6-6329-4e6c-9cf2-b51f272fd6b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.842424] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245126, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.883029] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 879.883029] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0131e-8a6f-242a-9908-b452f6de6e9c" [ 879.883029] env[63418]: _type = "HttpNfcLease" [ 879.883029] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 879.883029] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 879.883029] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d0131e-8a6f-242a-9908-b452f6de6e9c" [ 879.883029] env[63418]: _type = "HttpNfcLease" [ 879.883029] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 879.883029] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8fa87f0-e823-46fa-8bdb-fa0e19a6a58d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.898089] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52512558-a87e-444b-3b17-4b5c18bfb236/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 879.898089] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52512558-a87e-444b-3b17-4b5c18bfb236/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 879.993686] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-be9ceeb4-f1c6-401a-9841-983ace44a7bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.062909] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245130, 'name': ReconfigVM_Task, 'duration_secs': 0.526105} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.063230] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 5ab04ddd-4cd6-4330-a9d7-081290c27586/5ab04ddd-4cd6-4330-a9d7-081290c27586.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.063893] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62b99857-2854-46c2-bc4d-0b86505b264b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.079669] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 880.079669] env[63418]: value = "task-1245132" [ 880.079669] env[63418]: _type = "Task" [ 880.079669] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.096341] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245132, 'name': Rename_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.204806] env[63418]: INFO nova.compute.manager [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Took 32.10 seconds to build instance. [ 880.305237] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752bab6d-9c14-43d2-8f00-99cc3eafa3bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.313824] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d665cbc-5ce2-4075-ac95-1f778acccfb7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.350493] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d28edac-a1e8-4bbb-808e-b6c9e9d30af8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.360180] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245126, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.439202} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.362689] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/14d64c0d-64ca-4b9e-9590-07ca59045ade/14d64c0d-64ca-4b9e-9590-07ca59045ade.vmdk to [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.364482] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfaeceb3-5982-43f7-a03c-550d75fd94f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.367361] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f12a8a9-b435-48cc-a115-4935a41f9c1f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.382838] env[63418]: DEBUG nova.compute.provider_tree [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.402203] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.404384] env[63418]: DEBUG nova.scheduler.client.report [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 880.407732] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8cfaeae-8456-4880-bbdc-b653987920cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.424073] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.384s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.424759] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 880.427596] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.499s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.429041] env[63418]: INFO nova.compute.claims [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.438854] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 880.438854] env[63418]: value = "task-1245133" [ 880.438854] env[63418]: _type = "Task" [ 880.438854] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.449059] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.590236] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245132, 'name': Rename_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.707602] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0872e75e-13d9-4fae-9b82-35ee94066800 tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.616s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.936648] env[63418]: DEBUG nova.compute.utils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.938554] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 880.938732] env[63418]: DEBUG nova.network.neutron [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.953874] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.997139] env[63418]: DEBUG nova.policy [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2229b43c3ec044d797bc24ae311f6619', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ed6de22ab90f40698f31305b9c7f18c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.093224] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245132, 'name': Rename_Task, 'duration_secs': 0.77929} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.093573] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.093953] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c809202a-12e2-4ccb-863d-f5a8f53f1f95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.101349] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 881.101349] env[63418]: value = "task-1245134" [ 881.101349] env[63418]: _type = "Task" [ 881.101349] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.112458] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245134, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.296905] env[63418]: DEBUG nova.network.neutron [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Successfully created port: c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.444771] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 881.462347] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245133, 'name': ReconfigVM_Task, 'duration_secs': 0.856617} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.462928] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfigured VM instance instance-00000031 to attach disk [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca/d76a008c-9bd9-420b-873d-4f7d7f25b8ca.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.465978] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'encryption_options': None, 'encrypted': False, 'encryption_secret_uuid': None, 'disk_bus': None, 'encryption_format': None, 'boot_index': 0, 'device_name': '/dev/sda', 'device_type': 'disk', 'guest_format': None, 'image_id': 'c0dccf29-5e49-4a1e-b51f-d46e566b4772'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'attachment_id': '6c5afe76-b23f-4211-a936-38be25c5fc25', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268463', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'name': 'volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'd76a008c-9bd9-420b-873d-4f7d7f25b8ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'serial': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967'}, 'disk_bus': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'device_type': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=63418) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 881.466660] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 881.466762] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268463', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'name': 'volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'd76a008c-9bd9-420b-873d-4f7d7f25b8ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'serial': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 881.471876] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ac9302-bc6b-4b25-8d1f-c1426f56538e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.499072] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d61bf38-72b5-4dce-ab1b-09f2dba19e0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.529019] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967/volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.532773] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54f6643d-5d67-4c8e-9a7f-047d4589597f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.554878] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 881.554878] env[63418]: value = "task-1245135" [ 881.554878] env[63418]: _type = "Task" [ 881.554878] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.565432] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245135, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.614634] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245134, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.737907] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2a7fec-f06a-4a3d-92fe-1fcf4ddc009e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.746808] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26059fe1-eba2-4dab-9e73-7affa33a3e59 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.781638] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9b14ec-15de-4bd1-a8ad-566940c291e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.791014] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177dc3d9-72a3-4a5b-a7bc-fde0b912d8b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.806480] env[63418]: DEBUG nova.compute.provider_tree [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.941473] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "384b4ae7-ad55-4d44-bd10-3b82534c9703" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.941746] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.942035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "384b4ae7-ad55-4d44-bd10-3b82534c9703-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.942275] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.942546] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.945367] env[63418]: INFO nova.compute.manager [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Terminating instance [ 882.068351] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245135, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.113554] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245134, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.309584] env[63418]: DEBUG nova.scheduler.client.report [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 882.452117] env[63418]: DEBUG nova.compute.manager [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 882.452117] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.452117] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede6b493-fba1-4799-a8a1-4c1821ce552f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.459290] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 882.470566] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.470934] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83c7e294-553c-4db0-86df-9014c58a558d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.480455] env[63418]: DEBUG oslo_vmware.api [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 882.480455] env[63418]: value = "task-1245136" [ 882.480455] env[63418]: _type = "Task" [ 882.480455] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.485633] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.486522] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.486522] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.486522] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.486687] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.486722] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.486937] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.487122] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.487284] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.487492] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.487774] env[63418]: DEBUG nova.virt.hardware [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.488691] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16734c0-18ac-4977-b0e1-24968e6c18b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.500291] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf3d21a-d2c6-4100-8b97-8b4f79d1d443 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.504563] env[63418]: DEBUG oslo_vmware.api [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.565705] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245135, 'name': ReconfigVM_Task, 'duration_secs': 0.541091} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.566075] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfigured VM instance instance-00000031 to attach disk [datastore1] volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967/volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.571813] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-918512ad-8706-4252-a6e4-15e840c8baf6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.589392] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 882.589392] env[63418]: value = "task-1245137" [ 882.589392] env[63418]: _type = "Task" [ 882.589392] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.599540] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.614719] env[63418]: DEBUG oslo_vmware.api [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245134, 'name': PowerOnVM_Task, 'duration_secs': 1.26236} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.615020] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.615280] env[63418]: INFO nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Took 9.88 seconds to spawn the instance on the hypervisor. [ 882.615480] env[63418]: DEBUG nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.616346] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c473f6a-0316-4749-93cb-ef2079cf14ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.741780] env[63418]: DEBUG nova.compute.manager [req-7ade7697-5927-41d9-8ef6-6c826cc554e8 req-a5dec79e-54a6-4cf7-907a-a202539192c2 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Received event network-vif-plugged-c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 882.742066] env[63418]: DEBUG oslo_concurrency.lockutils [req-7ade7697-5927-41d9-8ef6-6c826cc554e8 req-a5dec79e-54a6-4cf7-907a-a202539192c2 service nova] Acquiring lock "586f51a0-90de-4f44-ac3e-758ceda9e316-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.742329] env[63418]: DEBUG oslo_concurrency.lockutils [req-7ade7697-5927-41d9-8ef6-6c826cc554e8 req-a5dec79e-54a6-4cf7-907a-a202539192c2 service nova] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.742511] env[63418]: DEBUG oslo_concurrency.lockutils [req-7ade7697-5927-41d9-8ef6-6c826cc554e8 req-a5dec79e-54a6-4cf7-907a-a202539192c2 service nova] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.742689] env[63418]: DEBUG nova.compute.manager [req-7ade7697-5927-41d9-8ef6-6c826cc554e8 req-a5dec79e-54a6-4cf7-907a-a202539192c2 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] No waiting events found dispatching network-vif-plugged-c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 882.742886] env[63418]: WARNING nova.compute.manager [req-7ade7697-5927-41d9-8ef6-6c826cc554e8 req-a5dec79e-54a6-4cf7-907a-a202539192c2 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Received unexpected event network-vif-plugged-c72f76ce-b5dd-4767-95d6-a97488e1bdd4 for instance with vm_state building and task_state spawning. [ 882.815708] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.816382] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 882.818980] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.651s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.819288] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.821366] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.329s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.821607] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.824194] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.133s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.824404] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.826525] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.393s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.827963] env[63418]: INFO nova.compute.claims [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.841982] env[63418]: DEBUG nova.network.neutron [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Successfully updated port: c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.855670] env[63418]: INFO nova.scheduler.client.report [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted allocations for instance dcc94934-d8d2-46d9-85fd-e17c11a7a4d7 [ 882.857914] env[63418]: INFO nova.scheduler.client.report [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Deleted allocations for instance e19a60cd-4ec7-48ad-9042-f19b75353364 [ 882.871088] env[63418]: INFO nova.scheduler.client.report [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocations for instance e2dcb8b8-e778-4202-8808-6a8535e2f1b4 [ 882.991363] env[63418]: DEBUG oslo_vmware.api [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245136, 'name': PowerOffVM_Task, 'duration_secs': 0.302307} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.991597] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.991868] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.992214] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f8987ae-6cbc-415f-989f-a2ac05050c6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.065268] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.065555] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.065733] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Deleting the datastore file [datastore1] 384b4ae7-ad55-4d44-bd10-3b82534c9703 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.066057] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec4d8f65-2855-4f3d-881f-98e4edfe9d44 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.073615] env[63418]: DEBUG oslo_vmware.api [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for the task: (returnval){ [ 883.073615] env[63418]: value = "task-1245139" [ 883.073615] env[63418]: _type = "Task" [ 883.073615] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.082465] env[63418]: DEBUG oslo_vmware.api [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.098907] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245137, 'name': ReconfigVM_Task, 'duration_secs': 0.233265} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.099307] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268463', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'name': 'volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'd76a008c-9bd9-420b-873d-4f7d7f25b8ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'serial': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 883.099958] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed66ea16-2beb-4ce5-8395-e74d4581c1e7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.107187] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 883.107187] env[63418]: value = "task-1245140" [ 883.107187] env[63418]: _type = "Task" [ 883.107187] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.116760] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245140, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.134659] env[63418]: INFO nova.compute.manager [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Took 27.54 seconds to build instance. [ 883.332751] env[63418]: DEBUG nova.compute.utils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.336444] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 883.336633] env[63418]: DEBUG nova.network.neutron [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.344322] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "refresh_cache-586f51a0-90de-4f44-ac3e-758ceda9e316" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.344485] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquired lock "refresh_cache-586f51a0-90de-4f44-ac3e-758ceda9e316" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.344630] env[63418]: DEBUG nova.network.neutron [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.365637] env[63418]: DEBUG oslo_concurrency.lockutils [None req-83d6d878-a3d2-4452-b040-4b050ac6f9a3 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "dcc94934-d8d2-46d9-85fd-e17c11a7a4d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.394s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.368897] env[63418]: DEBUG oslo_concurrency.lockutils [None req-096516c5-87a7-4a38-8e0a-71e118e4f76b tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "e19a60cd-4ec7-48ad-9042-f19b75353364" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.782s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.380454] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a323f121-13db-4943-b57b-831e4632e2c1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "e2dcb8b8-e778-4202-8808-6a8535e2f1b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.728s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.386092] env[63418]: DEBUG nova.policy [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ef8f62bd75a4d109a2e1881408754cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11fffb9badce4abeadce3ab70dff7d58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.584979] env[63418]: DEBUG oslo_vmware.api [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Task: {'id': task-1245139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135274} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.585375] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.585593] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.585784] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.585979] env[63418]: INFO nova.compute.manager [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Took 1.14 seconds to destroy the instance on the hypervisor. [ 883.586281] env[63418]: DEBUG oslo.service.loopingcall [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.586598] env[63418]: DEBUG nova.compute.manager [-] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 883.586736] env[63418]: DEBUG nova.network.neutron [-] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 883.620127] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245140, 'name': Rename_Task, 'duration_secs': 0.222091} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.620127] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.620127] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97b1e761-fbc0-4dfb-b6bd-93b05243bd1c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.629687] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 883.629687] env[63418]: value = "task-1245141" [ 883.629687] env[63418]: _type = "Task" [ 883.629687] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.639340] env[63418]: DEBUG oslo_concurrency.lockutils [None req-33977f61-e9bd-4afc-bca1-36d1966b40f2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.065s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.639651] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245141, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.748826] env[63418]: DEBUG nova.network.neutron [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Successfully created port: e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.838149] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 883.887757] env[63418]: DEBUG nova.network.neutron [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.095779] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbd087d-87ac-43aa-8de8-f622a57e272c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.104970] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc98d55-9675-494c-8cbd-1169d6ca6d0d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.144190] env[63418]: DEBUG nova.network.neutron [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Updating instance_info_cache with network_info: [{"id": "c72f76ce-b5dd-4767-95d6-a97488e1bdd4", "address": "fa:16:3e:45:5d:85", "network": {"id": "bfa75a01-2fc4-43d1-8117-02f3449f5e9f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1967670944-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed6de22ab90f40698f31305b9c7f18c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc72f76ce-b5", "ovs_interfaceid": "c72f76ce-b5dd-4767-95d6-a97488e1bdd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.150025] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2a71d4-a835-4af6-bae4-176aba0b946b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.164115] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245141, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.166149] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4a9c00-3504-45ed-b065-a7ba48e0bdc2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.181828] env[63418]: DEBUG nova.compute.provider_tree [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.517928] env[63418]: DEBUG nova.network.neutron [-] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.654047] env[63418]: DEBUG oslo_vmware.api [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245141, 'name': PowerOnVM_Task, 'duration_secs': 0.632662} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.654387] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.656833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Releasing lock "refresh_cache-586f51a0-90de-4f44-ac3e-758ceda9e316" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.657126] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Instance network_info: |[{"id": "c72f76ce-b5dd-4767-95d6-a97488e1bdd4", "address": "fa:16:3e:45:5d:85", "network": {"id": "bfa75a01-2fc4-43d1-8117-02f3449f5e9f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1967670944-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed6de22ab90f40698f31305b9c7f18c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc72f76ce-b5", "ovs_interfaceid": "c72f76ce-b5dd-4767-95d6-a97488e1bdd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 884.659807] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:5d:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68add7d6-c025-46fa-84d3-9c589adb63e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c72f76ce-b5dd-4767-95d6-a97488e1bdd4', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.667968] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Creating folder: Project (ed6de22ab90f40698f31305b9c7f18c3). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.668755] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-92ddaea2-4414-4c0a-af45-7ef7fe15fc50 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.685549] env[63418]: DEBUG nova.scheduler.client.report [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 884.690935] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Created folder: Project (ed6de22ab90f40698f31305b9c7f18c3) in parent group-v268354. [ 884.691209] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Creating folder: Instances. Parent ref: group-v268480. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.691471] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4232533e-4346-4c17-810c-5ae4bd6c64bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.704874] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Created folder: Instances in parent group-v268480. [ 884.704874] env[63418]: DEBUG oslo.service.loopingcall [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.705581] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.705581] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4466fc71-5068-4261-a26a-85e6d3c3b088 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.726459] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.726459] env[63418]: value = "task-1245145" [ 884.726459] env[63418]: _type = "Task" [ 884.726459] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.736047] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245145, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.759330] env[63418]: DEBUG nova.compute.manager [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 884.760291] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a0e911-206c-4403-8b1c-80f871fef795 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.854547] env[63418]: DEBUG nova.compute.manager [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Received event network-changed-c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 884.854899] env[63418]: DEBUG nova.compute.manager [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Refreshing instance network info cache due to event network-changed-c72f76ce-b5dd-4767-95d6-a97488e1bdd4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 884.855167] env[63418]: DEBUG oslo_concurrency.lockutils [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] Acquiring lock "refresh_cache-586f51a0-90de-4f44-ac3e-758ceda9e316" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.855378] env[63418]: DEBUG oslo_concurrency.lockutils [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] Acquired lock "refresh_cache-586f51a0-90de-4f44-ac3e-758ceda9e316" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.855616] env[63418]: DEBUG nova.network.neutron [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Refreshing network info cache for port c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.858010] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 884.880437] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.880785] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.880997] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.881278] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.881511] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.881727] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.882532] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.882532] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.882532] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.882684] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.882854] env[63418]: DEBUG nova.virt.hardware [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.884161] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82821b2d-2211-407e-b644-6671b408aaf2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.893970] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401ee27c-1993-498d-8de8-2c48e4e312f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.020417] env[63418]: INFO nova.compute.manager [-] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Took 1.43 seconds to deallocate network for instance. [ 885.171498] env[63418]: DEBUG nova.compute.manager [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 885.172892] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f6a5c2-d917-4c6a-872d-48c69a87e2f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.195364] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.199152] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 885.199578] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.416s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.201168] env[63418]: INFO nova.compute.claims [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.237939] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245145, 'name': CreateVM_Task, 'duration_secs': 0.388932} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.238125] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.238816] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.238990] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.239341] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.239608] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa28f581-7930-4160-b49b-5127f749d28d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.245596] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 885.245596] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523c3777-b99d-7f10-7805-24808ed72b12" [ 885.245596] env[63418]: _type = "Task" [ 885.245596] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.255484] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523c3777-b99d-7f10-7805-24808ed72b12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.279904] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aec246c7-3bb1-41e4-9a05-735d90a04d23 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 46.666s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.383698] env[63418]: DEBUG nova.network.neutron [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Successfully updated port: e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.527900] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.634383] env[63418]: DEBUG nova.network.neutron [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Updated VIF entry in instance network info cache for port c72f76ce-b5dd-4767-95d6-a97488e1bdd4. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.634781] env[63418]: DEBUG nova.network.neutron [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Updating instance_info_cache with network_info: [{"id": "c72f76ce-b5dd-4767-95d6-a97488e1bdd4", "address": "fa:16:3e:45:5d:85", "network": {"id": "bfa75a01-2fc4-43d1-8117-02f3449f5e9f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1967670944-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed6de22ab90f40698f31305b9c7f18c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc72f76ce-b5", "ovs_interfaceid": "c72f76ce-b5dd-4767-95d6-a97488e1bdd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.684925] env[63418]: INFO nova.compute.manager [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] instance snapshotting [ 885.687817] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1af7db-74fe-4838-a0e5-1abec5747cac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.713227] env[63418]: DEBUG nova.compute.utils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.716838] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 885.717030] env[63418]: DEBUG nova.network.neutron [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.719393] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138a8c36-dd1d-4695-b949-76b6223b3a69 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.757831] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523c3777-b99d-7f10-7805-24808ed72b12, 'name': SearchDatastore_Task, 'duration_secs': 0.010814} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.758229] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.758490] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.758745] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.758902] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.759161] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.759459] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd426ef9-dee7-4b51-b64c-93266b889aaa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.771252] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.771466] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.772769] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8a43851-a23e-4e18-b8e6-d3d67d2e7333 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.779412] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 885.779412] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ad75ff-2f42-23ff-7a51-3cf32108acd1" [ 885.779412] env[63418]: _type = "Task" [ 885.779412] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.789190] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ad75ff-2f42-23ff-7a51-3cf32108acd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.792454] env[63418]: DEBUG nova.policy [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b0f6e35d8494335b25867ab54fe1a3d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e074b295ca4d4e50bde88f6d37a7f98f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.887234] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.887234] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.887487] env[63418]: DEBUG nova.network.neutron [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.137216] env[63418]: DEBUG oslo_concurrency.lockutils [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] Releasing lock "refresh_cache-586f51a0-90de-4f44-ac3e-758ceda9e316" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.137514] env[63418]: DEBUG nova.compute.manager [req-48165573-d21c-4da2-ad0a-8abf92cac8d3 req-91295d44-dcc7-4782-a1fe-a98451f5d6f4 service nova] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Received event network-vif-deleted-606a5028-84ee-4542-aa24-d9a709985395 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 886.218177] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 886.232245] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 886.232599] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3424200a-4230-423b-bf61-59c80a1d0fa7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.241368] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 886.241368] env[63418]: value = "task-1245146" [ 886.241368] env[63418]: _type = "Task" [ 886.241368] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.251246] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245146, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.294062] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ad75ff-2f42-23ff-7a51-3cf32108acd1, 'name': SearchDatastore_Task, 'duration_secs': 0.021934} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.295540] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adb3d51c-6739-4d9a-b2b2-db04d701fb4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.301863] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 886.301863] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a2b223-701a-4f16-3d98-c12e075a016a" [ 886.301863] env[63418]: _type = "Task" [ 886.301863] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.318900] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a2b223-701a-4f16-3d98-c12e075a016a, 'name': SearchDatastore_Task, 'duration_secs': 0.01139} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.319216] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.319502] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 586f51a0-90de-4f44-ac3e-758ceda9e316/586f51a0-90de-4f44-ac3e-758ceda9e316.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.319875] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5b7aa1b-078a-4bd8-9fba-59711a6dfb0f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.332244] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 886.332244] env[63418]: value = "task-1245147" [ 886.332244] env[63418]: _type = "Task" [ 886.332244] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.333091] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.333091] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.333223] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.333739] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.333739] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.335698] env[63418]: DEBUG nova.network.neutron [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Successfully created port: 87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.338312] env[63418]: INFO nova.compute.manager [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Terminating instance [ 886.350564] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245147, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.427398] env[63418]: DEBUG nova.network.neutron [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.491474] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95eb1b27-0667-467c-8127-c99523dd6e12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.505599] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18997e4d-73ce-4c14-b40e-34356722b008 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.540437] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73db328f-7c60-4cc7-9414-e79028f22bd9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.549595] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afffd30-1385-4b44-bcd6-04b60dafda76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.567380] env[63418]: DEBUG nova.compute.provider_tree [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.753298] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245146, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.819986] env[63418]: DEBUG nova.network.neutron [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance_info_cache with network_info: [{"id": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "address": "fa:16:3e:72:aa:e2", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ff1dd4-2e", "ovs_interfaceid": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.845981] env[63418]: DEBUG nova.compute.manager [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 886.845981] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.849953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b5b0e5-7d89-4041-83a9-f1d48bf89f18 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.853612] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245147, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473241} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.854285] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 586f51a0-90de-4f44-ac3e-758ceda9e316/586f51a0-90de-4f44-ac3e-758ceda9e316.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.854595] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.855227] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8baebd5-59a0-4c08-8c68-fee78acffb6c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.860477] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.860632] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e233eca-afec-4498-8078-b91e546d3a8f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.863717] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 886.863717] env[63418]: value = "task-1245148" [ 886.863717] env[63418]: _type = "Task" [ 886.863717] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.870285] env[63418]: DEBUG oslo_vmware.api [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 886.870285] env[63418]: value = "task-1245149" [ 886.870285] env[63418]: _type = "Task" [ 886.870285] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.874053] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245148, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.886316] env[63418]: DEBUG oslo_vmware.api [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.888750] env[63418]: DEBUG nova.compute.manager [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Received event network-vif-plugged-e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 886.889046] env[63418]: DEBUG oslo_concurrency.lockutils [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] Acquiring lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.889273] env[63418]: DEBUG oslo_concurrency.lockutils [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.889445] env[63418]: DEBUG oslo_concurrency.lockutils [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.889636] env[63418]: DEBUG nova.compute.manager [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] No waiting events found dispatching network-vif-plugged-e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 886.889813] env[63418]: WARNING nova.compute.manager [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Received unexpected event network-vif-plugged-e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 for instance with vm_state building and task_state spawning. [ 886.889982] env[63418]: DEBUG nova.compute.manager [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Received event network-changed-e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 886.890163] env[63418]: DEBUG nova.compute.manager [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Refreshing instance network info cache due to event network-changed-e4ff1dd4-2e2f-42cb-b882-19c984a2eb28. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 886.890355] env[63418]: DEBUG oslo_concurrency.lockutils [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] Acquiring lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.071433] env[63418]: DEBUG nova.scheduler.client.report [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 887.228501] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 887.254278] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245146, 'name': CreateSnapshot_Task, 'duration_secs': 0.804888} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.256500] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.256724] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.256886] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.257087] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.257240] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.257402] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.257685] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.257862] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.258059] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.258341] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.258553] env[63418]: DEBUG nova.virt.hardware [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.258884] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 887.259678] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa2a379-0b13-4a50-88e7-5efe76ac3602 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.264041] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1728448-94dd-4d28-8e60-846ca023eb06 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.275309] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace8f060-ed05-4654-80cd-eb813acca350 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.322244] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.322621] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Instance network_info: |[{"id": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "address": "fa:16:3e:72:aa:e2", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ff1dd4-2e", "ovs_interfaceid": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 887.323228] env[63418]: DEBUG oslo_concurrency.lockutils [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] Acquired lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.323445] env[63418]: DEBUG nova.network.neutron [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Refreshing network info cache for port e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.324759] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:aa:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4ff1dd4-2e2f-42cb-b882-19c984a2eb28', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.332306] env[63418]: DEBUG oslo.service.loopingcall [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.335284] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.335831] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-696ea29e-b390-4abf-9527-7bd8d82069ff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.358767] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.358767] env[63418]: value = "task-1245150" [ 887.358767] env[63418]: _type = "Task" [ 887.358767] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.371464] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245150, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.377318] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245148, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071806} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.382031] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.384869] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea16364-679c-4b35-9a43-c06b335d0a79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.394302] env[63418]: DEBUG oslo_vmware.api [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245149, 'name': PowerOffVM_Task, 'duration_secs': 0.259935} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.403648] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.403890] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.413848] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 586f51a0-90de-4f44-ac3e-758ceda9e316/586f51a0-90de-4f44-ac3e-758ceda9e316.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.414212] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-424462ca-6f29-43e2-b961-2bf5caa1bdbe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.416382] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5ddde8b-752a-4d4a-b54d-974458764de6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.441300] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 887.441300] env[63418]: value = "task-1245152" [ 887.441300] env[63418]: _type = "Task" [ 887.441300] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.451062] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245152, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.519996] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.520190] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.520511] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Deleting the datastore file [datastore2] 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.520857] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89c5fe0c-0139-458f-a48f-1725dfcc207d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.528880] env[63418]: DEBUG oslo_vmware.api [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for the task: (returnval){ [ 887.528880] env[63418]: value = "task-1245153" [ 887.528880] env[63418]: _type = "Task" [ 887.528880] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.539688] env[63418]: DEBUG oslo_vmware.api [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245153, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.577480] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.578100] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 887.581192] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.860s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.581458] env[63418]: DEBUG nova.objects.instance [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'resources' on Instance uuid eaa13276-1fb1-47e2-ad1f-445bc9f4c98f {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.608021] env[63418]: DEBUG nova.network.neutron [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updated VIF entry in instance network info cache for port e4ff1dd4-2e2f-42cb-b882-19c984a2eb28. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.608442] env[63418]: DEBUG nova.network.neutron [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance_info_cache with network_info: [{"id": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "address": "fa:16:3e:72:aa:e2", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ff1dd4-2e", "ovs_interfaceid": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.791802] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 887.792271] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6c5f82b5-9f3c-408e-bf39-f72ac902e438 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.802207] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 887.802207] env[63418]: value = "task-1245154" [ 887.802207] env[63418]: _type = "Task" [ 887.802207] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.811310] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245154, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.869959] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245150, 'name': CreateVM_Task, 'duration_secs': 0.447746} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.870202] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.870918] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.871127] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.871474] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.871789] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d04e1469-ec04-48b0-b2b9-3fcbdacebe90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.877252] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 887.877252] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52740773-e02a-762a-754b-edd97539d72d" [ 887.877252] env[63418]: _type = "Task" [ 887.877252] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.887487] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52740773-e02a-762a-754b-edd97539d72d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.960764] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245152, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.039670] env[63418]: DEBUG oslo_vmware.api [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Task: {'id': task-1245153, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162121} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.039935] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.040141] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.040331] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.040507] env[63418]: INFO nova.compute.manager [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Took 1.20 seconds to destroy the instance on the hypervisor. [ 888.040749] env[63418]: DEBUG oslo.service.loopingcall [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.040943] env[63418]: DEBUG nova.compute.manager [-] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 888.041050] env[63418]: DEBUG nova.network.neutron [-] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.084476] env[63418]: DEBUG nova.compute.utils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.086223] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 888.086409] env[63418]: DEBUG nova.network.neutron [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.112219] env[63418]: DEBUG oslo_concurrency.lockutils [req-7edb8480-6a39-4f8a-aec3-150042720e54 req-94d8be9a-a280-4b84-ab52-8fd6d40458a8 service nova] Releasing lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.114651] env[63418]: DEBUG nova.compute.manager [req-946d1d7e-84c1-4f00-8cf9-b7fc45f7a831 req-52fa4f85-3699-46c7-a983-7bbe6ddf0835 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Received event network-vif-plugged-87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 888.114881] env[63418]: DEBUG oslo_concurrency.lockutils [req-946d1d7e-84c1-4f00-8cf9-b7fc45f7a831 req-52fa4f85-3699-46c7-a983-7bbe6ddf0835 service nova] Acquiring lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.115171] env[63418]: DEBUG oslo_concurrency.lockutils [req-946d1d7e-84c1-4f00-8cf9-b7fc45f7a831 req-52fa4f85-3699-46c7-a983-7bbe6ddf0835 service nova] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.115383] env[63418]: DEBUG oslo_concurrency.lockutils [req-946d1d7e-84c1-4f00-8cf9-b7fc45f7a831 req-52fa4f85-3699-46c7-a983-7bbe6ddf0835 service nova] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.115558] env[63418]: DEBUG nova.compute.manager [req-946d1d7e-84c1-4f00-8cf9-b7fc45f7a831 req-52fa4f85-3699-46c7-a983-7bbe6ddf0835 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] No waiting events found dispatching network-vif-plugged-87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 888.115743] env[63418]: WARNING nova.compute.manager [req-946d1d7e-84c1-4f00-8cf9-b7fc45f7a831 req-52fa4f85-3699-46c7-a983-7bbe6ddf0835 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Received unexpected event network-vif-plugged-87ac533a-e3c0-46b3-984b-b7631bdc3e22 for instance with vm_state building and task_state spawning. [ 888.166592] env[63418]: DEBUG nova.policy [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.314398] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245154, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.330318] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb327de0-56b9-40b0-b36a-29a1953d4eff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.339570] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5a092c-c90f-45ef-a6dc-2ea4326ecf61 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.377344] env[63418]: DEBUG nova.network.neutron [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Successfully updated port: 87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.379570] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee12052-9b95-4f6c-945a-c277f69c8114 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.397769] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48932d88-c218-412e-a2fc-7fdb153cf350 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.402825] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52740773-e02a-762a-754b-edd97539d72d, 'name': SearchDatastore_Task, 'duration_secs': 0.010677} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.403319] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.403659] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.403902] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.404112] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.404377] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.405437] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8ce7fa0-c389-47b0-8cae-547ae7e49690 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.417985] env[63418]: DEBUG nova.compute.provider_tree [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.427979] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.428207] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.429553] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fc1084e-41f8-473d-8424-ffd4a7993d22 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.437885] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 888.437885] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525485ec-87ae-ab18-3a24-ff7e9f30c866" [ 888.437885] env[63418]: _type = "Task" [ 888.437885] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.449818] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525485ec-87ae-ab18-3a24-ff7e9f30c866, 'name': SearchDatastore_Task, 'duration_secs': 0.010264} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.455173] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afa45155-e7c9-41a2-8a7b-5e32874882de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.463792] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245152, 'name': ReconfigVM_Task, 'duration_secs': 0.697597} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.465358] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 586f51a0-90de-4f44-ac3e-758ceda9e316/586f51a0-90de-4f44-ac3e-758ceda9e316.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.466536] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 888.466536] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52144966-71df-0a4f-c90a-c9c05ad449ed" [ 888.466536] env[63418]: _type = "Task" [ 888.466536] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.466865] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1441d330-e318-4c1e-a61d-940b2c1e8e7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.479475] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52144966-71df-0a4f-c90a-c9c05ad449ed, 'name': SearchDatastore_Task, 'duration_secs': 0.011232} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.479743] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.480048] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5/1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.480851] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c285b23-992d-47b2-975c-5f20ebcc8818 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.483766] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 888.483766] env[63418]: value = "task-1245155" [ 888.483766] env[63418]: _type = "Task" [ 888.483766] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.489491] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 888.489491] env[63418]: value = "task-1245156" [ 888.489491] env[63418]: _type = "Task" [ 888.489491] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.496346] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245155, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.503436] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.590437] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 888.802208] env[63418]: DEBUG nova.network.neutron [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Successfully created port: b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.816925] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245154, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.885334] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.885646] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.885859] env[63418]: DEBUG nova.network.neutron [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.914355] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52512558-a87e-444b-3b17-4b5c18bfb236/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 888.915434] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a29fe3-1f92-4afa-aa27-f6516c9ad3a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.924094] env[63418]: DEBUG nova.scheduler.client.report [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 888.928452] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52512558-a87e-444b-3b17-4b5c18bfb236/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 888.928678] env[63418]: ERROR oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52512558-a87e-444b-3b17-4b5c18bfb236/disk-0.vmdk due to incomplete transfer. [ 888.928956] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d762e597-9562-49d8-a921-aa73b3a4c721 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.939695] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52512558-a87e-444b-3b17-4b5c18bfb236/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 888.939999] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Uploaded image 8d2a987b-43f8-48c5-a6e9-e586f49f246f to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 888.943345] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 888.943664] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f7884abf-fd87-432c-996f-288b405af1d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.952591] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 888.952591] env[63418]: value = "task-1245157" [ 888.952591] env[63418]: _type = "Task" [ 888.952591] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.961925] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245157, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.996323] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245155, 'name': Rename_Task, 'duration_secs': 0.244979} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.996993] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.997268] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abcd5282-b85f-4dd4-b515-2d35a5cf4aab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.001860] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480147} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.002967] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5/1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.002967] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.002967] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-543bae2e-c5e7-44c4-a968-f95da656e1f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.008661] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 889.008661] env[63418]: value = "task-1245158" [ 889.008661] env[63418]: _type = "Task" [ 889.008661] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.013639] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 889.013639] env[63418]: value = "task-1245159" [ 889.013639] env[63418]: _type = "Task" [ 889.013639] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.021733] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.028223] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245159, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.036775] env[63418]: DEBUG nova.compute.manager [req-dbbaffdd-c30c-4850-96e3-652d29b7e9c9 req-8b4d76bb-c2a5-4d0f-897b-f5be6a821875 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Received event network-vif-deleted-8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 889.036873] env[63418]: INFO nova.compute.manager [req-dbbaffdd-c30c-4850-96e3-652d29b7e9c9 req-8b4d76bb-c2a5-4d0f-897b-f5be6a821875 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Neutron deleted interface 8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a; detaching it from the instance and deleting it from the info cache [ 889.037173] env[63418]: DEBUG nova.network.neutron [req-dbbaffdd-c30c-4850-96e3-652d29b7e9c9 req-8b4d76bb-c2a5-4d0f-897b-f5be6a821875 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.315411] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245154, 'name': CloneVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.421167] env[63418]: DEBUG nova.network.neutron [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.432394] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.435821] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.363s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.436985] env[63418]: INFO nova.compute.claims [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.458999] env[63418]: INFO nova.scheduler.client.report [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted allocations for instance eaa13276-1fb1-47e2-ad1f-445bc9f4c98f [ 889.470947] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245157, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.509741] env[63418]: DEBUG nova.network.neutron [-] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.530142] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245158, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.533607] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245159, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076174} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.534107] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.535181] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac451ac-22cf-4db7-91d5-1e703ad1ab1f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.540383] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c063aa96-730b-4ade-8d18-51298c238957 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.564890] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5/1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.567935] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71c88c86-d412-4462-935d-e612e6224fc1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.593345] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea9b170-fd91-4661-a3b6-ea8407c9c416 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.606878] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 889.611024] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 889.611024] env[63418]: value = "task-1245160" [ 889.611024] env[63418]: _type = "Task" [ 889.611024] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.620316] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245160, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.630081] env[63418]: DEBUG nova.compute.manager [req-dbbaffdd-c30c-4850-96e3-652d29b7e9c9 req-8b4d76bb-c2a5-4d0f-897b-f5be6a821875 service nova] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Detach interface failed, port_id=8f91f4e7-2ec8-466e-ad3e-1c4794c70e3a, reason: Instance 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 889.646352] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.647142] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.647142] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.647303] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.647441] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.647591] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.647816] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.648062] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.648237] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.648425] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.648661] env[63418]: DEBUG nova.virt.hardware [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.649682] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda70124-b454-4f0a-bc68-346cdc742fab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.659381] env[63418]: DEBUG nova.network.neutron [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Updating instance_info_cache with network_info: [{"id": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "address": "fa:16:3e:0b:52:29", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ac533a-e3", "ovs_interfaceid": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.661774] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be4a946-76f1-4dc9-8265-c9c661c085c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.815338] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245154, 'name': CloneVM_Task, 'duration_secs': 1.667179} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.815638] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Created linked-clone VM from snapshot [ 889.816465] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475314b3-875b-4665-8d8a-999eabb2e540 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.824531] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Uploading image c2b900b6-9949-4e61-8a79-99af7b0febe6 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 889.835663] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 889.835949] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0885a780-0554-457c-ac5d-ccff16b8b1e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.843303] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 889.843303] env[63418]: value = "task-1245161" [ 889.843303] env[63418]: _type = "Task" [ 889.843303] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.851501] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245161, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.968498] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245157, 'name': Destroy_Task, 'duration_secs': 1.006201} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.968724] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroyed the VM [ 889.968967] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 889.969244] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-255734a4-6581-4ed1-b21e-74dfcd0306f1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.975095] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d92c8488-014a-47c8-b325-86f98800838f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "eaa13276-1fb1-47e2-ad1f-445bc9f4c98f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.192s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.977360] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 889.977360] env[63418]: value = "task-1245162" [ 889.977360] env[63418]: _type = "Task" [ 889.977360] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.990925] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245162, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.016924] env[63418]: INFO nova.compute.manager [-] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Took 1.98 seconds to deallocate network for instance. [ 890.025546] env[63418]: DEBUG oslo_vmware.api [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245158, 'name': PowerOnVM_Task, 'duration_secs': 1.003442} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.029084] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.029204] env[63418]: INFO nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Took 7.57 seconds to spawn the instance on the hypervisor. [ 890.029396] env[63418]: DEBUG nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 890.030613] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8c3fbc-86ba-43a9-a996-39b6bb1f0336 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.125675] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245160, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.167366] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.167725] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Instance network_info: |[{"id": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "address": "fa:16:3e:0b:52:29", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ac533a-e3", "ovs_interfaceid": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 890.168188] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:52:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b36c5ae6-c344-4bd1-8239-29128e2bbfbf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87ac533a-e3c0-46b3-984b-b7631bdc3e22', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.178603] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Creating folder: Project (e074b295ca4d4e50bde88f6d37a7f98f). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.178936] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c23b9f8e-d383-4b3b-9584-1f677add6f9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.193206] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Created folder: Project (e074b295ca4d4e50bde88f6d37a7f98f) in parent group-v268354. [ 890.193206] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Creating folder: Instances. Parent ref: group-v268486. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.193528] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c016525-c12a-4480-b069-643cd7b63cf4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.204402] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Created folder: Instances in parent group-v268486. [ 890.204719] env[63418]: DEBUG oslo.service.loopingcall [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.204931] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.205166] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b80c25e-bd6b-47ef-be10-bbc578e9de23 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.226454] env[63418]: DEBUG nova.compute.manager [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Received event network-changed-87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 890.226454] env[63418]: DEBUG nova.compute.manager [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Refreshing instance network info cache due to event network-changed-87ac533a-e3c0-46b3-984b-b7631bdc3e22. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 890.226454] env[63418]: DEBUG oslo_concurrency.lockutils [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] Acquiring lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.226454] env[63418]: DEBUG oslo_concurrency.lockutils [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] Acquired lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.226454] env[63418]: DEBUG nova.network.neutron [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Refreshing network info cache for port 87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.234995] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.234995] env[63418]: value = "task-1245165" [ 890.234995] env[63418]: _type = "Task" [ 890.234995] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.247282] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245165, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.355646] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245161, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.492329] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245162, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.531411] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.555224] env[63418]: INFO nova.compute.manager [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Took 29.09 seconds to build instance. [ 890.621963] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245160, 'name': ReconfigVM_Task, 'duration_secs': 0.651275} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.624875] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5/1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.625982] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-455aaf69-2bb9-4990-8cbb-052fe9059fc5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.640139] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 890.640139] env[63418]: value = "task-1245166" [ 890.640139] env[63418]: _type = "Task" [ 890.640139] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.649581] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245166, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.671489] env[63418]: DEBUG nova.network.neutron [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Successfully updated port: b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.696632] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3739bc4d-b91a-4fc4-818c-c055c57825f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.707957] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fdbd46-fd5a-4dd9-aa00-4c9d526c0675 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.754687] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58763e57-a5e4-4f30-9bb7-57d473dbf925 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.763775] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245165, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.768794] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad8a4ad-e2f4-4fb2-9d27-8c6043727616 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.785633] env[63418]: DEBUG nova.compute.provider_tree [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.857550] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245161, 'name': Destroy_Task, 'duration_secs': 0.593442} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.858286] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Destroyed the VM [ 890.858599] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 890.858892] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-dee07fa4-c5b8-4dac-a40f-3c4e981f51de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.869740] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 890.869740] env[63418]: value = "task-1245167" [ 890.869740] env[63418]: _type = "Task" [ 890.869740] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.877728] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245167, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.967087] env[63418]: DEBUG nova.network.neutron [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Updated VIF entry in instance network info cache for port 87ac533a-e3c0-46b3-984b-b7631bdc3e22. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.967483] env[63418]: DEBUG nova.network.neutron [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Updating instance_info_cache with network_info: [{"id": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "address": "fa:16:3e:0b:52:29", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ac533a-e3", "ovs_interfaceid": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.991485] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245162, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.056541] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7f57d81e-c012-48ce-ad8c-30dce333f045 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.604s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.084785] env[63418]: DEBUG nova.compute.manager [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Received event network-vif-plugged-b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 891.085028] env[63418]: DEBUG oslo_concurrency.lockutils [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] Acquiring lock "bf091892-fefb-49dd-9416-708a06b35798-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.085356] env[63418]: DEBUG oslo_concurrency.lockutils [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] Lock "bf091892-fefb-49dd-9416-708a06b35798-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.087371] env[63418]: DEBUG oslo_concurrency.lockutils [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] Lock "bf091892-fefb-49dd-9416-708a06b35798-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.087371] env[63418]: DEBUG nova.compute.manager [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] No waiting events found dispatching network-vif-plugged-b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 891.087371] env[63418]: WARNING nova.compute.manager [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Received unexpected event network-vif-plugged-b1cdc19c-4c4d-4568-ade7-c0b97713632b for instance with vm_state building and task_state spawning. [ 891.087371] env[63418]: DEBUG nova.compute.manager [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Received event network-changed-b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 891.087371] env[63418]: DEBUG nova.compute.manager [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Refreshing instance network info cache due to event network-changed-b1cdc19c-4c4d-4568-ade7-c0b97713632b. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 891.087371] env[63418]: DEBUG oslo_concurrency.lockutils [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] Acquiring lock "refresh_cache-bf091892-fefb-49dd-9416-708a06b35798" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.087371] env[63418]: DEBUG oslo_concurrency.lockutils [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] Acquired lock "refresh_cache-bf091892-fefb-49dd-9416-708a06b35798" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.088019] env[63418]: DEBUG nova.network.neutron [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Refreshing network info cache for port b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.150075] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245166, 'name': Rename_Task, 'duration_secs': 0.201635} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.150375] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.150629] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8dfdd82-5c23-4722-94e1-b3a6869f0502 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.157306] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 891.157306] env[63418]: value = "task-1245168" [ 891.157306] env[63418]: _type = "Task" [ 891.157306] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.169407] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.175076] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-bf091892-fefb-49dd-9416-708a06b35798" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.260749] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245165, 'name': CreateVM_Task, 'duration_secs': 0.547957} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.260928] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.261647] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.261814] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.262193] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.262464] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67859028-6a50-4cc8-a023-c6ca2dad8502 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.269300] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 891.269300] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5282474e-ff65-3bab-62fe-0a3c0cff738e" [ 891.269300] env[63418]: _type = "Task" [ 891.269300] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.279030] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5282474e-ff65-3bab-62fe-0a3c0cff738e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.293175] env[63418]: DEBUG nova.scheduler.client.report [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 891.379737] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245167, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.470306] env[63418]: DEBUG oslo_concurrency.lockutils [req-ef5e4973-faa6-4bd0-93bd-0a89d5575b87 req-7d8bd800-a7de-4a8e-b429-851231187362 service nova] Releasing lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.492303] env[63418]: DEBUG oslo_vmware.api [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245162, 'name': RemoveSnapshot_Task, 'duration_secs': 1.128564} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.495021] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 891.495021] env[63418]: INFO nova.compute.manager [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 16.32 seconds to snapshot the instance on the hypervisor. [ 891.634996] env[63418]: DEBUG nova.network.neutron [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.667597] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245168, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.755623] env[63418]: DEBUG nova.network.neutron [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.785482] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5282474e-ff65-3bab-62fe-0a3c0cff738e, 'name': SearchDatastore_Task, 'duration_secs': 0.011963} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.785838] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.786098] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.786346] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.786513] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.786708] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.786991] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2132b210-6a43-434c-a1e8-75a15d68d138 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.799140] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.799717] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 891.802717] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.802717] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.804519] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.373s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.804519] env[63418]: DEBUG nova.objects.instance [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lazy-loading 'resources' on Instance uuid 1f213c5b-de27-48d6-a1f8-fdf9b77359fd {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.805247] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-781f107b-4d6c-4cb1-9188-5b634448883f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.812945] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 891.812945] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d4dec4-b2d7-70ab-0788-c1bd5a40a070" [ 891.812945] env[63418]: _type = "Task" [ 891.812945] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.823016] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d4dec4-b2d7-70ab-0788-c1bd5a40a070, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.881427] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245167, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.053544] env[63418]: DEBUG nova.compute.manager [None req-e1d5ffa4-aa4e-48ab-9d45-d8995c7480c1 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Found 2 images (rotation: 2) {{(pid=63418) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 892.169108] env[63418]: DEBUG oslo_vmware.api [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245168, 'name': PowerOnVM_Task, 'duration_secs': 0.821528} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.169367] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.169691] env[63418]: INFO nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Took 7.31 seconds to spawn the instance on the hypervisor. [ 892.169889] env[63418]: DEBUG nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 892.170683] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cc7fa9-d49a-486b-8159-863ba3f56cf2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.262961] env[63418]: DEBUG oslo_concurrency.lockutils [req-7677e3b4-dcff-4619-abef-78cd813254ec req-3d61ed8e-5418-47fb-93f8-9b0e79479148 service nova] Releasing lock "refresh_cache-bf091892-fefb-49dd-9416-708a06b35798" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.262961] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-bf091892-fefb-49dd-9416-708a06b35798" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.262961] env[63418]: DEBUG nova.network.neutron [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.306535] env[63418]: DEBUG nova.compute.utils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.308379] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 892.308679] env[63418]: DEBUG nova.network.neutron [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.338372] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "586f51a0-90de-4f44-ac3e-758ceda9e316" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.338689] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.339032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "586f51a0-90de-4f44-ac3e-758ceda9e316-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.339372] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.339666] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.342544] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d4dec4-b2d7-70ab-0788-c1bd5a40a070, 'name': SearchDatastore_Task, 'duration_secs': 0.010057} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.343168] env[63418]: INFO nova.compute.manager [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Terminating instance [ 892.350774] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7bcddc6-2731-4a37-be91-fdf94a9ac25b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.362024] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 892.362024] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bfcf43-a7ef-e0cd-6855-fa5c099a87bf" [ 892.362024] env[63418]: _type = "Task" [ 892.362024] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.376675] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bfcf43-a7ef-e0cd-6855-fa5c099a87bf, 'name': SearchDatastore_Task, 'duration_secs': 0.011339} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.378072] env[63418]: DEBUG nova.policy [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea507bad11c3406d880ba47d08a047c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd445600834dd4c7e8022349ee993f3ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.383218] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.383218] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.383218] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c65a4d64-26cd-4124-8f25-406ceac1b795 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.388166] env[63418]: DEBUG oslo_vmware.api [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245167, 'name': RemoveSnapshot_Task, 'duration_secs': 1.049912} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.388874] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 892.394291] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 892.394291] env[63418]: value = "task-1245169" [ 892.394291] env[63418]: _type = "Task" [ 892.394291] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.409185] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245169, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.544459] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04745758-1ece-4ff0-b362-d2c295d19714 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.554185] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdb6ca7-da16-4c11-a4b6-31ba91b30dcc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.591076] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa7edac-ed73-447a-9006-0564e88d0872 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.600211] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f883c1-7ee7-424c-8971-222596793a2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.618835] env[63418]: DEBUG nova.compute.provider_tree [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.689665] env[63418]: INFO nova.compute.manager [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Took 30.79 seconds to build instance. [ 892.709052] env[63418]: DEBUG nova.compute.manager [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 892.710039] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99837cc4-54ae-4139-8809-ea4242e103cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.812580] env[63418]: DEBUG nova.network.neutron [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.821715] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 892.858941] env[63418]: DEBUG nova.compute.manager [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 892.859218] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 892.861665] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b6b92c-6053-4088-b4d2-f71e8fffad40 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.876945] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.882201] env[63418]: DEBUG nova.network.neutron [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Successfully created port: 65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.885538] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a96434d6-c880-4c4a-8466-99a40002ef54 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.893681] env[63418]: WARNING nova.compute.manager [None req-839a81a3-ee28-4bd9-aef8-d43976701b5f tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Image not found during snapshot: nova.exception.ImageNotFound: Image c2b900b6-9949-4e61-8a79-99af7b0febe6 could not be found. [ 892.904642] env[63418]: DEBUG oslo_vmware.api [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 892.904642] env[63418]: value = "task-1245170" [ 892.904642] env[63418]: _type = "Task" [ 892.904642] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.916231] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245169, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.924441] env[63418]: DEBUG oslo_vmware.api [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.073672] env[63418]: DEBUG nova.network.neutron [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Updating instance_info_cache with network_info: [{"id": "b1cdc19c-4c4d-4568-ade7-c0b97713632b", "address": "fa:16:3e:57:15:fa", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1cdc19c-4c", "ovs_interfaceid": "b1cdc19c-4c4d-4568-ade7-c0b97713632b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.122490] env[63418]: DEBUG nova.scheduler.client.report [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 893.165945] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.166216] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.192816] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c38fa7f-eb89-4caf-911d-ebd50705260e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.303s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.223944] env[63418]: INFO nova.compute.manager [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] instance snapshotting [ 893.223944] env[63418]: DEBUG nova.objects.instance [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'flavor' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.415748] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245169, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577323} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.416453] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.416746] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.417023] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e14b00c2-971c-4955-a6e8-bdce51f2714d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.425762] env[63418]: DEBUG oslo_vmware.api [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245170, 'name': PowerOffVM_Task, 'duration_secs': 0.373096} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.426619] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.426670] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 893.426953] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-495971d7-8b5c-42c3-a619-f8fd7cd03914 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.432589] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 893.432589] env[63418]: value = "task-1245171" [ 893.432589] env[63418]: _type = "Task" [ 893.432589] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.443730] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.503243] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 893.503511] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 893.503701] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Deleting the datastore file [datastore2] 586f51a0-90de-4f44-ac3e-758ceda9e316 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.504988] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-296b9b01-b6ed-48e0-90fd-fc1297eecc76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.514281] env[63418]: DEBUG oslo_vmware.api [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for the task: (returnval){ [ 893.514281] env[63418]: value = "task-1245173" [ 893.514281] env[63418]: _type = "Task" [ 893.514281] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.524340] env[63418]: DEBUG oslo_vmware.api [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.577455] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-bf091892-fefb-49dd-9416-708a06b35798" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.578178] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Instance network_info: |[{"id": "b1cdc19c-4c4d-4568-ade7-c0b97713632b", "address": "fa:16:3e:57:15:fa", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1cdc19c-4c", "ovs_interfaceid": "b1cdc19c-4c4d-4568-ade7-c0b97713632b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 893.578617] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:15:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1cdc19c-4c4d-4568-ade7-c0b97713632b', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.592131] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating folder: Project (4dac4885b78c414bbe379918a0d8cf61). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.592576] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44bc3770-88b7-44f6-92c0-653615df2f9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.606305] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created folder: Project (4dac4885b78c414bbe379918a0d8cf61) in parent group-v268354. [ 893.606620] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating folder: Instances. Parent ref: group-v268489. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.606913] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-257480cd-42b3-474f-886f-8d4684543216 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.618135] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created folder: Instances in parent group-v268489. [ 893.618406] env[63418]: DEBUG oslo.service.loopingcall [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.618637] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf091892-fefb-49dd-9416-708a06b35798] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 893.618870] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c94f6a9-0b00-4822-9954-b17472e30dd6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.635945] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.638829] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.996s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.640528] env[63418]: INFO nova.compute.claims [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.650889] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.650889] env[63418]: value = "task-1245176" [ 893.650889] env[63418]: _type = "Task" [ 893.650889] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.661763] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245176, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.663074] env[63418]: INFO nova.scheduler.client.report [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted allocations for instance 1f213c5b-de27-48d6-a1f8-fdf9b77359fd [ 893.670466] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 893.729639] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b5fca2-d810-419b-b09e-f36d873071d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.751173] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ae3e81-c7e0-444e-add8-6658dc08639d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.795053] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "5ab04ddd-4cd6-4330-a9d7-081290c27586" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.795404] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.795636] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "5ab04ddd-4cd6-4330-a9d7-081290c27586-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.795830] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.796099] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.798917] env[63418]: INFO nova.compute.manager [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Terminating instance [ 893.835150] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 893.860493] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.860732] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.860888] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.861080] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.861228] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.861374] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.861581] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.861736] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.862061] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.862135] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.863682] env[63418]: DEBUG nova.virt.hardware [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.863682] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a9bb45-a59e-4fef-84c3-d58bae326c48 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.871516] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd68c64e-8d54-484b-9f8f-ba1ccc500f8f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.943259] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245171, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120395} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.943580] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.945030] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8382d8f2-8f4c-411f-9399-21760b569668 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.967558] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.967884] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8aa4141f-d60a-400b-9c68-b699a3917a2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.990190] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 893.990190] env[63418]: value = "task-1245177" [ 893.990190] env[63418]: _type = "Task" [ 893.990190] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.999313] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.026016] env[63418]: DEBUG oslo_vmware.api [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Task: {'id': task-1245173, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15178} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.026306] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 894.026520] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 894.026706] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 894.026918] env[63418]: INFO nova.compute.manager [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Took 1.17 seconds to destroy the instance on the hypervisor. [ 894.027316] env[63418]: DEBUG oslo.service.loopingcall [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.027564] env[63418]: DEBUG nova.compute.manager [-] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 894.027683] env[63418]: DEBUG nova.network.neutron [-] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 894.161765] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245176, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.177684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bed6360b-7713-4e6c-87fa-70f433c0ba3e tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "1f213c5b-de27-48d6-a1f8-fdf9b77359fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.717s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.198528] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.265349] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 894.265664] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-69ccc635-2f0f-4687-892b-44e349ce682c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.275842] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 894.275842] env[63418]: value = "task-1245178" [ 894.275842] env[63418]: _type = "Task" [ 894.275842] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.292977] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245178, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.303021] env[63418]: DEBUG nova.compute.manager [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 894.303208] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.304380] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6892e91d-93a0-4237-a4ef-ec41ce81df0b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.313770] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.314049] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24b24694-b571-445b-bc3b-a13cf0a76310 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.324107] env[63418]: DEBUG oslo_vmware.api [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 894.324107] env[63418]: value = "task-1245179" [ 894.324107] env[63418]: _type = "Task" [ 894.324107] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.336682] env[63418]: DEBUG oslo_vmware.api [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.369732] env[63418]: DEBUG nova.compute.manager [req-e7edeecf-3954-471d-b95c-e1a266f7367b req-5d8c0b49-a8ab-4b9c-a994-841fa145a6e3 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Received event network-vif-deleted-c72f76ce-b5dd-4767-95d6-a97488e1bdd4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 894.369958] env[63418]: INFO nova.compute.manager [req-e7edeecf-3954-471d-b95c-e1a266f7367b req-5d8c0b49-a8ab-4b9c-a994-841fa145a6e3 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Neutron deleted interface c72f76ce-b5dd-4767-95d6-a97488e1bdd4; detaching it from the instance and deleting it from the info cache [ 894.370221] env[63418]: DEBUG nova.network.neutron [req-e7edeecf-3954-471d-b95c-e1a266f7367b req-5d8c0b49-a8ab-4b9c-a994-841fa145a6e3 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.506430] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245177, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.632237] env[63418]: DEBUG nova.network.neutron [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Successfully updated port: 65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.665965] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245176, 'name': CreateVM_Task, 'duration_secs': 0.829429} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.666620] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf091892-fefb-49dd-9416-708a06b35798] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.667821] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.668589] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.669063] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.672167] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5d832e5-db44-4ef0-9628-60bfc75783e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.678373] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 894.678373] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52448dc9-9c2c-7686-77c8-75110d080509" [ 894.678373] env[63418]: _type = "Task" [ 894.678373] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.693066] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52448dc9-9c2c-7686-77c8-75110d080509, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.788072] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245178, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.831663] env[63418]: DEBUG nova.network.neutron [-] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.835447] env[63418]: DEBUG oslo_vmware.api [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245179, 'name': PowerOffVM_Task, 'duration_secs': 0.283} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.835928] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.836141] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.836406] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da5d8e57-291c-487e-a901-4450588300ef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.853738] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9ba5ba-be20-42f5-8795-2bf87c40e63b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.863641] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f39fe94-b94f-4c53-a97a-92098f1d1b79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.898054] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6dc07927-fd64-45c5-b4e9-16ce7586a77e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.900553] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdacb12d-3ea1-45ca-b2cb-ede0818f2c7d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.911056] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6495cd-50c6-4219-a398-f11c6488386c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.918327] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08856ed8-ba23-4011-97ba-76ce2d35fb59 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.929234] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 894.929594] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 894.929712] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleting the datastore file [datastore1] 5ab04ddd-4cd6-4330-a9d7-081290c27586 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.930726] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec8f8030-b472-4b7e-9ebe-6031683c986e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.941600] env[63418]: DEBUG nova.compute.provider_tree [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.949891] env[63418]: DEBUG nova.compute.manager [req-e7edeecf-3954-471d-b95c-e1a266f7367b req-5d8c0b49-a8ab-4b9c-a994-841fa145a6e3 service nova] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Detach interface failed, port_id=c72f76ce-b5dd-4767-95d6-a97488e1bdd4, reason: Instance 586f51a0-90de-4f44-ac3e-758ceda9e316 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 894.952197] env[63418]: DEBUG oslo_vmware.api [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 894.952197] env[63418]: value = "task-1245181" [ 894.952197] env[63418]: _type = "Task" [ 894.952197] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.961676] env[63418]: DEBUG oslo_vmware.api [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.000789] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245177, 'name': ReconfigVM_Task, 'duration_secs': 0.671867} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.001089] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Reconfigured VM instance instance-00000050 to attach disk [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.001713] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2019ece5-6fd9-45c8-ad0b-e120ec84ae41 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.009229] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 895.009229] env[63418]: value = "task-1245182" [ 895.009229] env[63418]: _type = "Task" [ 895.009229] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.022611] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245182, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.135293] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "refresh_cache-dcb52a03-8f82-4b33-8dc6-3924140db8d3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.135350] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "refresh_cache-dcb52a03-8f82-4b33-8dc6-3924140db8d3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.135521] env[63418]: DEBUG nova.network.neutron [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.189364] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52448dc9-9c2c-7686-77c8-75110d080509, 'name': SearchDatastore_Task, 'duration_secs': 0.009741} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.189697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.189958] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.190219] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.190373] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.190609] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.190886] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0796e011-247f-4c95-bb99-065b7c0d2f94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.200090] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.201032] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.201193] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd15b793-0f50-4136-929f-2bf5b328ea9e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.207830] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 895.207830] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d83453-ea7a-cd74-36b2-cb0c248c702c" [ 895.207830] env[63418]: _type = "Task" [ 895.207830] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.217046] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d83453-ea7a-cd74-36b2-cb0c248c702c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.287860] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245178, 'name': CreateSnapshot_Task, 'duration_secs': 0.82925} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.288170] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 895.288930] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ae30fc-2a31-4b3a-bdee-4f39f40da8cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.336367] env[63418]: INFO nova.compute.manager [-] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Took 1.31 seconds to deallocate network for instance. [ 895.448075] env[63418]: DEBUG nova.scheduler.client.report [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 895.463286] env[63418]: DEBUG oslo_vmware.api [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146005} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.464232] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.464446] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.464625] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.464801] env[63418]: INFO nova.compute.manager [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Took 1.16 seconds to destroy the instance on the hypervisor. [ 895.465098] env[63418]: DEBUG oslo.service.loopingcall [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.465382] env[63418]: DEBUG nova.compute.manager [-] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 895.465382] env[63418]: DEBUG nova.network.neutron [-] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.473674] env[63418]: DEBUG nova.compute.manager [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Stashing vm_state: active {{(pid=63418) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 895.522843] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245182, 'name': Rename_Task, 'duration_secs': 0.231658} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.523393] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.523492] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59e517b6-b970-4f5f-a354-6f7f5404d163 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.531932] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 895.531932] env[63418]: value = "task-1245183" [ 895.531932] env[63418]: _type = "Task" [ 895.531932] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.540430] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.669590] env[63418]: DEBUG nova.network.neutron [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.719133] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d83453-ea7a-cd74-36b2-cb0c248c702c, 'name': SearchDatastore_Task, 'duration_secs': 0.009856} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.720057] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b578b19-6dbe-453b-9de9-bcf3c61f3719 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.726860] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 895.726860] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523fffad-3244-b366-6b21-c222fffa4259" [ 895.726860] env[63418]: _type = "Task" [ 895.726860] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.739956] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523fffad-3244-b366-6b21-c222fffa4259, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.808020] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 895.810632] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-58104eee-1c81-4b85-8f40-e27df00c8d5a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.822007] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 895.822007] env[63418]: value = "task-1245184" [ 895.822007] env[63418]: _type = "Task" [ 895.822007] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.832668] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245184, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.842852] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.850232] env[63418]: DEBUG nova.network.neutron [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Updating instance_info_cache with network_info: [{"id": "65da165e-6d27-4c6a-9c82-0186992d1648", "address": "fa:16:3e:87:02:77", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65da165e-6d", "ovs_interfaceid": "65da165e-6d27-4c6a-9c82-0186992d1648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.949894] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.950597] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 895.953353] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.426s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.953589] env[63418]: DEBUG nova.objects.instance [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lazy-loading 'resources' on Instance uuid 384b4ae7-ad55-4d44-bd10-3b82534c9703 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.995492] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.043396] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245183, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.237973] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523fffad-3244-b366-6b21-c222fffa4259, 'name': SearchDatastore_Task, 'duration_secs': 0.010621} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.238264] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.238529] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] bf091892-fefb-49dd-9416-708a06b35798/bf091892-fefb-49dd-9416-708a06b35798.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.238789] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-488ef78a-7874-4989-8c41-ac3821ecc357 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.245568] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 896.245568] env[63418]: value = "task-1245185" [ 896.245568] env[63418]: _type = "Task" [ 896.245568] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.249524] env[63418]: DEBUG nova.network.neutron [-] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.255939] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.334022] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245184, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.353230] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "refresh_cache-dcb52a03-8f82-4b33-8dc6-3924140db8d3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.353748] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Instance network_info: |[{"id": "65da165e-6d27-4c6a-9c82-0186992d1648", "address": "fa:16:3e:87:02:77", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65da165e-6d", "ovs_interfaceid": "65da165e-6d27-4c6a-9c82-0186992d1648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 896.354288] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:02:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65da165e-6d27-4c6a-9c82-0186992d1648', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.362694] env[63418]: DEBUG oslo.service.loopingcall [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.363277] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.363277] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59e1ac24-c141-4bb1-b103-05fc011a49b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.384983] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.384983] env[63418]: value = "task-1245186" [ 896.384983] env[63418]: _type = "Task" [ 896.384983] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.394440] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245186, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.457945] env[63418]: DEBUG nova.compute.utils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.467347] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 896.467927] env[63418]: DEBUG nova.network.neutron [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.543299] env[63418]: DEBUG oslo_vmware.api [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245183, 'name': PowerOnVM_Task, 'duration_secs': 0.997219} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.543653] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.544014] env[63418]: INFO nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Took 9.32 seconds to spawn the instance on the hypervisor. [ 896.544131] env[63418]: DEBUG nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 896.545255] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d81ae3b-8fa5-4aef-95a1-b3633da41e76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.554336] env[63418]: DEBUG nova.compute.manager [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Received event network-vif-plugged-65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 896.554570] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] Acquiring lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.554810] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.554983] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.555183] env[63418]: DEBUG nova.compute.manager [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] No waiting events found dispatching network-vif-plugged-65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 896.555390] env[63418]: WARNING nova.compute.manager [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Received unexpected event network-vif-plugged-65da165e-6d27-4c6a-9c82-0186992d1648 for instance with vm_state building and task_state spawning. [ 896.555563] env[63418]: DEBUG nova.compute.manager [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Received event network-changed-65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 896.555829] env[63418]: DEBUG nova.compute.manager [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Refreshing instance network info cache due to event network-changed-65da165e-6d27-4c6a-9c82-0186992d1648. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 896.555976] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] Acquiring lock "refresh_cache-dcb52a03-8f82-4b33-8dc6-3924140db8d3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.556156] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] Acquired lock "refresh_cache-dcb52a03-8f82-4b33-8dc6-3924140db8d3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.556327] env[63418]: DEBUG nova.network.neutron [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Refreshing network info cache for port 65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.559252] env[63418]: DEBUG nova.policy [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ff64612530b451fb41100a5aa601be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d52f1fceb24234a8d967038b43c857', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.748024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e05f381-604f-48dc-a671-d87aa5501572 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.755760] env[63418]: INFO nova.compute.manager [-] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Took 1.29 seconds to deallocate network for instance. [ 896.772432] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b97053-5286-4ff2-90e7-6ec8a8ddfd60 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.776385] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245185, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458132} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.777172] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] bf091892-fefb-49dd-9416-708a06b35798/bf091892-fefb-49dd-9416-708a06b35798.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.777604] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.778449] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9beab16-3a30-4268-9443-1f7cbd38b020 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.811791] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb9574a-6960-4d68-a048-ccb912348e93 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.815310] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 896.815310] env[63418]: value = "task-1245187" [ 896.815310] env[63418]: _type = "Task" [ 896.815310] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.823222] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61991982-b69d-41ad-bc6d-001a16e4d56f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.830543] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245187, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.850712] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245184, 'name': CloneVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.851017] env[63418]: DEBUG nova.compute.provider_tree [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.875883] env[63418]: DEBUG nova.network.neutron [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Successfully created port: de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.895382] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245186, 'name': CreateVM_Task, 'duration_secs': 0.482974} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.895558] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.896278] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.896851] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.896851] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.898825] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6f362f6-2f7e-4d4f-860c-f0f10f8e74ef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.902589] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 896.902589] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a9a071-7ce5-e13a-d78c-290ddcdc1e64" [ 896.902589] env[63418]: _type = "Task" [ 896.902589] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.911357] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a9a071-7ce5-e13a-d78c-290ddcdc1e64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.967762] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 897.081917] env[63418]: INFO nova.compute.manager [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Took 24.67 seconds to build instance. [ 897.278592] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.306287] env[63418]: DEBUG nova.network.neutron [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Updated VIF entry in instance network info cache for port 65da165e-6d27-4c6a-9c82-0186992d1648. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.306659] env[63418]: DEBUG nova.network.neutron [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Updating instance_info_cache with network_info: [{"id": "65da165e-6d27-4c6a-9c82-0186992d1648", "address": "fa:16:3e:87:02:77", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65da165e-6d", "ovs_interfaceid": "65da165e-6d27-4c6a-9c82-0186992d1648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.325937] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245187, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213162} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.326210] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.327046] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1bbff2-69b5-4d8c-b229-def26a41932b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.350695] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] bf091892-fefb-49dd-9416-708a06b35798/bf091892-fefb-49dd-9416-708a06b35798.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.351299] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91f6c538-0241-41b3-a8c6-dbea5070215b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.375025] env[63418]: DEBUG nova.scheduler.client.report [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 897.378270] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245184, 'name': CloneVM_Task, 'duration_secs': 1.315644} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.379913] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Created linked-clone VM from snapshot [ 897.383185] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28aaa838-57b6-4244-8ab7-99e63e3bafe1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.387358] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 897.387358] env[63418]: value = "task-1245188" [ 897.387358] env[63418]: _type = "Task" [ 897.387358] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.395020] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Uploading image 9644ff2f-d0b1-47da-bc3d-36ffab2efcb5 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 897.403142] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245188, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.413789] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a9a071-7ce5-e13a-d78c-290ddcdc1e64, 'name': SearchDatastore_Task, 'duration_secs': 0.010699} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.414220] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.414495] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.414818] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.414988] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.415199] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.415515] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-869d92d7-ca14-425e-80b2-341ec7317cf6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.425521] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 897.425521] env[63418]: value = "vm-268493" [ 897.425521] env[63418]: _type = "VirtualMachine" [ 897.425521] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 897.425955] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8d6e45f1-279f-48cb-b347-27fcefd99230 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.428711] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.428899] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.429963] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7947ba43-6381-451c-8afb-02eae9f5de17 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.436470] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease: (returnval){ [ 897.436470] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52900e2f-5929-487c-334d-2899a0762dc7" [ 897.436470] env[63418]: _type = "HttpNfcLease" [ 897.436470] env[63418]: } obtained for exporting VM: (result){ [ 897.436470] env[63418]: value = "vm-268493" [ 897.436470] env[63418]: _type = "VirtualMachine" [ 897.436470] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 897.436470] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the lease: (returnval){ [ 897.436470] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52900e2f-5929-487c-334d-2899a0762dc7" [ 897.436470] env[63418]: _type = "HttpNfcLease" [ 897.436470] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 897.437263] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 897.437263] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525ecd4b-247c-1c65-0f40-5ec189e98464" [ 897.437263] env[63418]: _type = "Task" [ 897.437263] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.451889] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.451889] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52900e2f-5929-487c-334d-2899a0762dc7" [ 897.451889] env[63418]: _type = "HttpNfcLease" [ 897.451889] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 897.451889] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525ecd4b-247c-1c65-0f40-5ec189e98464, 'name': SearchDatastore_Task, 'duration_secs': 0.009838} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.452424] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beba328e-7204-4c98-9cbb-7b9004a2df4c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.457766] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 897.457766] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e6f17e-81b1-aec8-bae3-24bc6cb74fa4" [ 897.457766] env[63418]: _type = "Task" [ 897.457766] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.466832] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e6f17e-81b1-aec8-bae3-24bc6cb74fa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.582567] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d9707514-0f77-4e94-9dc7-5e91e532c150 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.178s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.809574] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] Releasing lock "refresh_cache-dcb52a03-8f82-4b33-8dc6-3924140db8d3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.809816] env[63418]: DEBUG nova.compute.manager [req-b11c722e-481e-404e-94a5-7071fb537b58 req-93362ef0-a257-49a4-a69f-e6f53f4b6b45 service nova] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Received event network-vif-deleted-ed4100a3-20c0-49ae-8fdd-778787b060aa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 897.880424] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.927s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.883880] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.352s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.883880] env[63418]: DEBUG nova.objects.instance [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lazy-loading 'resources' on Instance uuid 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.901741] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.911159] env[63418]: INFO nova.scheduler.client.report [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Deleted allocations for instance 384b4ae7-ad55-4d44-bd10-3b82534c9703 [ 897.948445] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.948445] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52900e2f-5929-487c-334d-2899a0762dc7" [ 897.948445] env[63418]: _type = "HttpNfcLease" [ 897.948445] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 897.948445] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 897.948445] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52900e2f-5929-487c-334d-2899a0762dc7" [ 897.948445] env[63418]: _type = "HttpNfcLease" [ 897.948445] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 897.948445] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6287b148-03d5-495d-a10b-75793d3022a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.956967] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ec972f-577b-b8e7-813b-e92df2aadc54/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 897.957413] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ec972f-577b-b8e7-813b-e92df2aadc54/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 898.020532] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 898.028935] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e6f17e-81b1-aec8-bae3-24bc6cb74fa4, 'name': SearchDatastore_Task, 'duration_secs': 0.009629} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.029608] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.029941] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] dcb52a03-8f82-4b33-8dc6-3924140db8d3/dcb52a03-8f82-4b33-8dc6-3924140db8d3.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.030277] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b902d001-7c99-48d9-992c-549c5ec39418 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.039807] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 898.039807] env[63418]: value = "task-1245190" [ 898.039807] env[63418]: _type = "Task" [ 898.039807] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.049289] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.060266] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.060484] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.060642] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.060825] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.060975] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.061142] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.061358] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.061521] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.061693] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.061859] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.062080] env[63418]: DEBUG nova.virt.hardware [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.062380] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8001ec42-1b85-4e74-8f65-e0c26855e83e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.064831] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835791d3-9811-46c2-9fb9-921125557377 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.077594] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67bfd7d6-34ad-4b2e-8fb8-2bea23566e64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.270123] env[63418]: INFO nova.compute.manager [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Rescuing [ 898.271025] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.271127] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.271243] env[63418]: DEBUG nova.network.neutron [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.403100] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245188, 'name': ReconfigVM_Task, 'duration_secs': 0.79665} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.403444] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Reconfigured VM instance instance-00000051 to attach disk [datastore2] bf091892-fefb-49dd-9416-708a06b35798/bf091892-fefb-49dd-9416-708a06b35798.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.404340] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c5b46c5-df8a-4b50-ba00-750a31711294 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.420842] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 898.420842] env[63418]: value = "task-1245191" [ 898.420842] env[63418]: _type = "Task" [ 898.420842] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.426187] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a468547d-3d84-4258-a315-1ef649dfba4a tempest-ServerPasswordTestJSON-880377299 tempest-ServerPasswordTestJSON-880377299-project-member] Lock "384b4ae7-ad55-4d44-bd10-3b82534c9703" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.484s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.437407] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245191, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.552296] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494851} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.552625] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] dcb52a03-8f82-4b33-8dc6-3924140db8d3/dcb52a03-8f82-4b33-8dc6-3924140db8d3.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.552844] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.553774] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5f1b954-0961-4799-b96e-807a1cbb680f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.563547] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 898.563547] env[63418]: value = "task-1245192" [ 898.563547] env[63418]: _type = "Task" [ 898.563547] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.574902] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245192, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.633974] env[63418]: DEBUG nova.compute.manager [req-ecc9c66d-b4d1-4c5d-aeac-a248f2cbe507 req-deff1933-8cf2-450c-9cf1-68e3f1110f80 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Received event network-vif-plugged-de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 898.634392] env[63418]: DEBUG oslo_concurrency.lockutils [req-ecc9c66d-b4d1-4c5d-aeac-a248f2cbe507 req-deff1933-8cf2-450c-9cf1-68e3f1110f80 service nova] Acquiring lock "054bdcde-55a6-4f86-a230-92b2115e47b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.634685] env[63418]: DEBUG oslo_concurrency.lockutils [req-ecc9c66d-b4d1-4c5d-aeac-a248f2cbe507 req-deff1933-8cf2-450c-9cf1-68e3f1110f80 service nova] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.634902] env[63418]: DEBUG oslo_concurrency.lockutils [req-ecc9c66d-b4d1-4c5d-aeac-a248f2cbe507 req-deff1933-8cf2-450c-9cf1-68e3f1110f80 service nova] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.635222] env[63418]: DEBUG nova.compute.manager [req-ecc9c66d-b4d1-4c5d-aeac-a248f2cbe507 req-deff1933-8cf2-450c-9cf1-68e3f1110f80 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] No waiting events found dispatching network-vif-plugged-de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 898.636082] env[63418]: WARNING nova.compute.manager [req-ecc9c66d-b4d1-4c5d-aeac-a248f2cbe507 req-deff1933-8cf2-450c-9cf1-68e3f1110f80 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Received unexpected event network-vif-plugged-de942ff3-dc1e-4603-9794-16ca7223fcb2 for instance with vm_state building and task_state spawning. [ 898.663526] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8112e5da-5254-4bce-8e5d-9ca3dd8f89c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.673108] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0b025d-9eca-4a3c-91ce-dc3b30e417b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.707636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c423fb-7012-4f19-a097-0d7cdd5c4386 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.716991] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9418cef1-ef8d-40bf-a8e2-2a3d4d1e3ca3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.732188] env[63418]: DEBUG nova.compute.provider_tree [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.935223] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245191, 'name': Rename_Task, 'duration_secs': 0.202472} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.935989] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.936366] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13a99983-57bb-4989-8be3-2bbe52d54f3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.946275] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 898.946275] env[63418]: value = "task-1245193" [ 898.946275] env[63418]: _type = "Task" [ 898.946275] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.960227] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.084503] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245192, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106103} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.096097] env[63418]: DEBUG nova.network.neutron [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Successfully updated port: de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.099550] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.099955] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0c29d9-8830-4f91-8a99-d012aeac3aed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.126331] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] dcb52a03-8f82-4b33-8dc6-3924140db8d3/dcb52a03-8f82-4b33-8dc6-3924140db8d3.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.127622] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-992e7d83-a53f-43e5-bdc6-307caf136f47 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.154679] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 899.154679] env[63418]: value = "task-1245194" [ 899.154679] env[63418]: _type = "Task" [ 899.154679] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.167582] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245194, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.235563] env[63418]: DEBUG nova.scheduler.client.report [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 899.250462] env[63418]: DEBUG nova.network.neutron [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Updating instance_info_cache with network_info: [{"id": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "address": "fa:16:3e:0b:52:29", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ac533a-e3", "ovs_interfaceid": "87ac533a-e3c0-46b3-984b-b7631bdc3e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.459111] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245193, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.599835] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-054bdcde-55a6-4f86-a230-92b2115e47b0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.600051] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-054bdcde-55a6-4f86-a230-92b2115e47b0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.600284] env[63418]: DEBUG nova.network.neutron [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.666911] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245194, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.741158] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.743996] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.545s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.746093] env[63418]: INFO nova.compute.claims [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.753868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "refresh_cache-bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.767373] env[63418]: INFO nova.scheduler.client.report [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Deleted allocations for instance 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171 [ 899.959267] env[63418]: DEBUG oslo_vmware.api [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245193, 'name': PowerOnVM_Task, 'duration_secs': 0.868441} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.959683] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.960148] env[63418]: INFO nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Took 10.35 seconds to spawn the instance on the hypervisor. [ 899.960625] env[63418]: DEBUG nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 899.961678] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3d68cc-3ee3-4a4e-bca9-3057f19897db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.155145] env[63418]: DEBUG nova.network.neutron [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.172327] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245194, 'name': ReconfigVM_Task, 'duration_secs': 0.602501} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.172645] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Reconfigured VM instance instance-00000052 to attach disk [datastore2] dcb52a03-8f82-4b33-8dc6-3924140db8d3/dcb52a03-8f82-4b33-8dc6-3924140db8d3.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.173390] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a208b472-d2b3-495a-90a4-87f458b90036 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.180776] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 900.180776] env[63418]: value = "task-1245195" [ 900.180776] env[63418]: _type = "Task" [ 900.180776] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.197281] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245195, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.281095] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6c739e0d-1682-48a4-8e26-66001cacad27 tempest-VolumesAdminNegativeTest-195519345 tempest-VolumesAdminNegativeTest-195519345-project-member] Lock "3e7b9a60-9e4b-46e3-adf9-7337dc7d2171" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.947s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.427881] env[63418]: DEBUG nova.network.neutron [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Updating instance_info_cache with network_info: [{"id": "de942ff3-dc1e-4603-9794-16ca7223fcb2", "address": "fa:16:3e:ff:1b:63", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde942ff3-dc", "ovs_interfaceid": "de942ff3-dc1e-4603-9794-16ca7223fcb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.482858] env[63418]: INFO nova.compute.manager [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Took 25.71 seconds to build instance. [ 900.694073] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245195, 'name': Rename_Task, 'duration_secs': 0.209443} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.694509] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.694787] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74ffa0bf-a8bb-48d1-9338-5fa01c20c67b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.702052] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 900.702052] env[63418]: value = "task-1245196" [ 900.702052] env[63418]: _type = "Task" [ 900.702052] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.711718] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.816720] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.817075] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76350db5-49fa-4177-bdcb-f2181da02b67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.824506] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 900.824506] env[63418]: value = "task-1245197" [ 900.824506] env[63418]: _type = "Task" [ 900.824506] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.837413] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245197, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.934473] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-054bdcde-55a6-4f86-a230-92b2115e47b0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.934767] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Instance network_info: |[{"id": "de942ff3-dc1e-4603-9794-16ca7223fcb2", "address": "fa:16:3e:ff:1b:63", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde942ff3-dc", "ovs_interfaceid": "de942ff3-dc1e-4603-9794-16ca7223fcb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 900.936761] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:1b:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de942ff3-dc1e-4603-9794-16ca7223fcb2', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.944950] env[63418]: DEBUG oslo.service.loopingcall [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.945270] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.945456] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bbd02eb-f41c-4139-81ee-6f5188644d42 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.968695] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.968695] env[63418]: value = "task-1245198" [ 900.968695] env[63418]: _type = "Task" [ 900.968695] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.979042] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245198, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.984830] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5549e645-0a55-4279-aeab-1151d2fe6cac tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "bf091892-fefb-49dd-9416-708a06b35798" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.232s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.002372] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74d0ad4-6abf-4418-86fa-d7713272ecd5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.010212] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a043e8-e06f-4693-97c0-9583e6705cf2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.045984] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b709d9c-098e-4578-936b-073582fdc1db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.050406] env[63418]: DEBUG nova.compute.manager [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Received event network-changed-de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 901.050664] env[63418]: DEBUG nova.compute.manager [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Refreshing instance network info cache due to event network-changed-de942ff3-dc1e-4603-9794-16ca7223fcb2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 901.050888] env[63418]: DEBUG oslo_concurrency.lockutils [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] Acquiring lock "refresh_cache-054bdcde-55a6-4f86-a230-92b2115e47b0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.051046] env[63418]: DEBUG oslo_concurrency.lockutils [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] Acquired lock "refresh_cache-054bdcde-55a6-4f86-a230-92b2115e47b0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.051215] env[63418]: DEBUG nova.network.neutron [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Refreshing network info cache for port de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.062527] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519c7dc9-55fb-484e-8be6-9e1436fc4109 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.079365] env[63418]: DEBUG nova.compute.provider_tree [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.213217] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245196, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.337235] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245197, 'name': PowerOffVM_Task, 'duration_secs': 0.252042} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.337584] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.338548] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385f0195-03b9-4a66-b6f6-a5889d0d7ad2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.359306] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201e4ca0-c630-455d-a534-bbae0697ae7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.397743] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.398273] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d57acf4e-7afb-4be4-873e-20d5751c7a04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.407159] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 901.407159] env[63418]: value = "task-1245199" [ 901.407159] env[63418]: _type = "Task" [ 901.407159] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.417954] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 901.417954] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.417954] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.417954] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.417954] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.418322] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2753e5b-0dc7-49e0-a29e-ea12442baac4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.428445] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.428757] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.429635] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9dcbbc6-ce81-4237-9b6f-ebce54107a63 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.436425] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 901.436425] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5299b941-d584-9481-3f6b-7271761a82cb" [ 901.436425] env[63418]: _type = "Task" [ 901.436425] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.445520] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5299b941-d584-9481-3f6b-7271761a82cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.483192] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245198, 'name': CreateVM_Task, 'duration_secs': 0.473536} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.484132] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.485077] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.485650] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.486279] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.486892] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd62baf4-2f10-4446-9714-42f25d48dde0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.493194] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 901.493194] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b1fb47-c3a7-d812-c185-d888addd500a" [ 901.493194] env[63418]: _type = "Task" [ 901.493194] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.503264] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b1fb47-c3a7-d812-c185-d888addd500a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.583297] env[63418]: DEBUG nova.scheduler.client.report [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 901.715646] env[63418]: DEBUG oslo_vmware.api [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245196, 'name': PowerOnVM_Task, 'duration_secs': 0.741368} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.716044] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.716168] env[63418]: INFO nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Took 7.88 seconds to spawn the instance on the hypervisor. [ 901.716337] env[63418]: DEBUG nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 901.717447] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f2208c-4bcb-49b4-9865-9f74dad19c4f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.948737] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5299b941-d584-9481-3f6b-7271761a82cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009891} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.949593] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a6a597b-59aa-4924-b305-5fc3e65859a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.955647] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 901.955647] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52484f3e-a48f-da9a-a088-5d332eb55167" [ 901.955647] env[63418]: _type = "Task" [ 901.955647] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.964316] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52484f3e-a48f-da9a-a088-5d332eb55167, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.979132] env[63418]: DEBUG nova.network.neutron [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Updated VIF entry in instance network info cache for port de942ff3-dc1e-4603-9794-16ca7223fcb2. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.979332] env[63418]: DEBUG nova.network.neutron [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Updating instance_info_cache with network_info: [{"id": "de942ff3-dc1e-4603-9794-16ca7223fcb2", "address": "fa:16:3e:ff:1b:63", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde942ff3-dc", "ovs_interfaceid": "de942ff3-dc1e-4603-9794-16ca7223fcb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.011448] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b1fb47-c3a7-d812-c185-d888addd500a, 'name': SearchDatastore_Task, 'duration_secs': 0.010071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.011761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.012014] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.012245] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.090357] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.090940] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 902.093870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.251s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.094305] env[63418]: DEBUG nova.objects.instance [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lazy-loading 'resources' on Instance uuid 586f51a0-90de-4f44-ac3e-758ceda9e316 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.237079] env[63418]: INFO nova.compute.manager [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Took 26.19 seconds to build instance. [ 902.467920] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52484f3e-a48f-da9a-a088-5d332eb55167, 'name': SearchDatastore_Task, 'duration_secs': 0.009883} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.468819] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.469208] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk. {{(pid=63418) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 902.469723] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.470016] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.470304] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f85b8d4-f86e-4e02-8e28-ab1441e3c8d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.472715] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f838fe89-54e1-4db4-addf-46bab7f4ae28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.482017] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 902.482017] env[63418]: value = "task-1245200" [ 902.482017] env[63418]: _type = "Task" [ 902.482017] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.482653] env[63418]: DEBUG oslo_concurrency.lockutils [req-8b1d8f2d-5cd3-4f7e-99b6-361527e7fe3a req-2386cc11-1bf9-43d9-9f68-041e68dc85f6 service nova] Releasing lock "refresh_cache-054bdcde-55a6-4f86-a230-92b2115e47b0" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.483453] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.483837] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.488805] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25230d3a-633c-4308-8e6f-b4765dbd17ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.497571] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.499199] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 902.499199] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d71fa7-a9c6-1605-ed43-c622b1284198" [ 902.499199] env[63418]: _type = "Task" [ 902.499199] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.509714] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d71fa7-a9c6-1605-ed43-c622b1284198, 'name': SearchDatastore_Task, 'duration_secs': 0.008965} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.510953] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-142a9f8d-5474-459a-ae51-2a978b9fc2ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.517616] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 902.517616] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5297471a-4975-d0c2-6d28-16d03e204b88" [ 902.517616] env[63418]: _type = "Task" [ 902.517616] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.527328] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5297471a-4975-d0c2-6d28-16d03e204b88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.600192] env[63418]: DEBUG nova.compute.utils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.603393] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 902.603677] env[63418]: DEBUG nova.network.neutron [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 902.674740] env[63418]: DEBUG nova.policy [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.747580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c18627dd-f892-4da8-bfdc-ad8a3cf5eaed tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.706s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.844823] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab8047a-84eb-45dd-b892-ea8411cf4ce6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.853923] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b528790-93aa-41aa-b964-c2f62fec5bd2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.892953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01a5577-9383-40db-954f-d5746abd5765 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.902404] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb50c0ea-e5f7-4b65-86fe-7a90df1ee91c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.923437] env[63418]: DEBUG nova.compute.provider_tree [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.993943] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490206} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.994246] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk. [ 902.995776] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f237d51f-e2c0-4362-a473-47f56e5fe7d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.026566] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.026566] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80fc6ef8-0cf0-4612-9577-da998358dcac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.050462] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5297471a-4975-d0c2-6d28-16d03e204b88, 'name': SearchDatastore_Task, 'duration_secs': 0.010068} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.050462] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.050822] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 054bdcde-55a6-4f86-a230-92b2115e47b0/054bdcde-55a6-4f86-a230-92b2115e47b0.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.050898] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 903.050898] env[63418]: value = "task-1245201" [ 903.050898] env[63418]: _type = "Task" [ 903.050898] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.051051] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3128177d-87d1-4bb7-9c37-cd6d83b04ef4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.061918] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245201, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.064039] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 903.064039] env[63418]: value = "task-1245202" [ 903.064039] env[63418]: _type = "Task" [ 903.064039] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.072195] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245202, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.105332] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 903.115585] env[63418]: DEBUG nova.network.neutron [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Successfully created port: 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.306826] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "14832e8e-3b4a-462e-84bb-d5a1e541d329" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.307024] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.429817] env[63418]: DEBUG nova.scheduler.client.report [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 903.566932] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245201, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.577192] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245202, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485535} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.577935] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 054bdcde-55a6-4f86-a230-92b2115e47b0/054bdcde-55a6-4f86-a230-92b2115e47b0.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.578104] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.578517] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ed57745-0591-4614-b4ed-48eb18907066 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.586762] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 903.586762] env[63418]: value = "task-1245203" [ 903.586762] env[63418]: _type = "Task" [ 903.586762] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.598163] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245203, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.809907] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 903.832085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.832479] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.833101] env[63418]: INFO nova.compute.manager [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Shelving [ 903.935570] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.841s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.939162] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.943s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.962979] env[63418]: INFO nova.scheduler.client.report [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Deleted allocations for instance 586f51a0-90de-4f44-ac3e-758ceda9e316 [ 904.066169] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245201, 'name': ReconfigVM_Task, 'duration_secs': 0.533618} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.066169] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Reconfigured VM instance instance-00000050 to attach disk [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.066964] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7a5072-0255-4d0b-82e3-ec58a1cf13e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.100391] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6df8d21a-5bc7-4728-8c0a-069bed979bdd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.114874] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 904.121032] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245203, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093701} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.122873] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.123409] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 904.123409] env[63418]: value = "task-1245204" [ 904.123409] env[63418]: _type = "Task" [ 904.123409] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.124359] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302a7d94-7f04-42ad-a07c-bda4300d4c05 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.146792] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.155811] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 054bdcde-55a6-4f86-a230-92b2115e47b0/054bdcde-55a6-4f86-a230-92b2115e47b0.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.158225] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.158472] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.158633] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.158820] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.158967] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.159131] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.159346] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.159510] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.159679] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.159843] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.160028] env[63418]: DEBUG nova.virt.hardware [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.160639] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31b89ce6-a2b1-45fc-98ee-0cc612102fe4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.175299] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90873c4-c92f-42cc-87c7-ca996a65ab2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.184898] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fad2655-249b-4a33-8c8f-54b76dde481d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.188932] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 904.188932] env[63418]: value = "task-1245205" [ 904.188932] env[63418]: _type = "Task" [ 904.188932] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.205702] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.333245] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.444707] env[63418]: INFO nova.compute.claims [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.474215] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c7a53d36-e1a9-4dc4-a1a0-179d10cef681 tempest-ServersNegativeTestMultiTenantJSON-1147750862 tempest-ServersNegativeTestMultiTenantJSON-1147750862-project-member] Lock "586f51a0-90de-4f44-ac3e-758ceda9e316" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.135s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.638049] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245204, 'name': ReconfigVM_Task, 'duration_secs': 0.437337} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.638432] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.638720] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd5a5f3e-9807-4ee1-91ee-af5844dd30a8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.645890] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 904.645890] env[63418]: value = "task-1245206" [ 904.645890] env[63418]: _type = "Task" [ 904.645890] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.656980] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.700058] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245205, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.740567] env[63418]: DEBUG nova.compute.manager [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 904.741533] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a91caca-12d4-429e-9a31-80f84246dbf6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.848771] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.849277] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-825f49f3-c462-45ef-8a2b-0ba939ca3f19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.859869] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 904.859869] env[63418]: value = "task-1245207" [ 904.859869] env[63418]: _type = "Task" [ 904.859869] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.869403] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245207, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.878586] env[63418]: DEBUG nova.compute.manager [req-ba444258-d393-4edc-9eeb-a02f835b1cde req-0caeca11-1b33-464b-91dd-d676d6d27428 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-vif-plugged-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 904.878586] env[63418]: DEBUG oslo_concurrency.lockutils [req-ba444258-d393-4edc-9eeb-a02f835b1cde req-0caeca11-1b33-464b-91dd-d676d6d27428 service nova] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.878917] env[63418]: DEBUG oslo_concurrency.lockutils [req-ba444258-d393-4edc-9eeb-a02f835b1cde req-0caeca11-1b33-464b-91dd-d676d6d27428 service nova] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.878917] env[63418]: DEBUG oslo_concurrency.lockutils [req-ba444258-d393-4edc-9eeb-a02f835b1cde req-0caeca11-1b33-464b-91dd-d676d6d27428 service nova] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.879048] env[63418]: DEBUG nova.compute.manager [req-ba444258-d393-4edc-9eeb-a02f835b1cde req-0caeca11-1b33-464b-91dd-d676d6d27428 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] No waiting events found dispatching network-vif-plugged-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 904.879262] env[63418]: WARNING nova.compute.manager [req-ba444258-d393-4edc-9eeb-a02f835b1cde req-0caeca11-1b33-464b-91dd-d676d6d27428 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received unexpected event network-vif-plugged-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 for instance with vm_state building and task_state spawning. [ 904.951433] env[63418]: INFO nova.compute.resource_tracker [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating resource usage from migration 64ccaa38-9e84-42b5-8d37-6092250c1d58 [ 905.147803] env[63418]: DEBUG nova.network.neutron [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Successfully updated port: 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.163235] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.179275] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4dc9fa-5ba7-4c1b-8252-363fb073b512 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.187287] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9fa7fa-dab2-436b-b887-46491f141f12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.222503] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb8647d-3aff-43e9-a58c-fc098219b1fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.232021] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245205, 'name': ReconfigVM_Task, 'duration_secs': 0.55121} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.232021] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 054bdcde-55a6-4f86-a230-92b2115e47b0/054bdcde-55a6-4f86-a230-92b2115e47b0.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.232021] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b54cc29-a2ee-411e-83b4-866bccdf8843 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.235927] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb6574b-81e0-4ff6-b56b-133f53f2f0e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.244159] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 905.244159] env[63418]: value = "task-1245208" [ 905.244159] env[63418]: _type = "Task" [ 905.244159] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.255855] env[63418]: INFO nova.compute.manager [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] instance snapshotting [ 905.258000] env[63418]: DEBUG nova.compute.provider_tree [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.262282] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245208, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.264040] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2ce22f-8d03-46a4-a365-a1ea14a2ced6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.284700] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cc3b3f-a8f7-4950-8929-0ea195c161e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.369872] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245207, 'name': PowerOffVM_Task, 'duration_secs': 0.22891} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.370759] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.371457] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f414f6-31b3-468e-9fa3-fbd67a6cd61c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.390665] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d5a7e6-b0ae-4efc-bcf2-2576c1ac19d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.605313] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ec972f-577b-b8e7-813b-e92df2aadc54/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 905.605313] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857556e8-5abe-4ffe-9026-4ce92822818d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.610956] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ec972f-577b-b8e7-813b-e92df2aadc54/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 905.611142] env[63418]: ERROR oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ec972f-577b-b8e7-813b-e92df2aadc54/disk-0.vmdk due to incomplete transfer. [ 905.611359] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d007f054-103e-467f-97ec-392a638a4244 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.618505] env[63418]: DEBUG oslo_vmware.rw_handles [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ec972f-577b-b8e7-813b-e92df2aadc54/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 905.618706] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Uploaded image 9644ff2f-d0b1-47da-bc3d-36ffab2efcb5 to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 905.621789] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 905.622057] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a50414a1-076f-4fd4-9b4f-ccd6d58cb975 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.627442] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 905.627442] env[63418]: value = "task-1245209" [ 905.627442] env[63418]: _type = "Task" [ 905.627442] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.635761] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245209, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.657992] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.658137] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.658286] env[63418]: DEBUG nova.network.neutron [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.660112] env[63418]: DEBUG oslo_vmware.api [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245206, 'name': PowerOnVM_Task, 'duration_secs': 0.608383} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.660112] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.665023] env[63418]: DEBUG nova.compute.manager [None req-2b59e380-5c75-4580-8260-086ed8e95910 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 905.665023] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f178c11d-09c0-44ec-9add-e56599891e3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.755870] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245208, 'name': Rename_Task, 'duration_secs': 0.194244} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.756049] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.756307] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ffb1652b-e9a5-4a39-8566-1674430e0af8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.761789] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 905.761789] env[63418]: value = "task-1245210" [ 905.761789] env[63418]: _type = "Task" [ 905.761789] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.766346] env[63418]: DEBUG nova.scheduler.client.report [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 905.775877] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.796693] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.797036] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b76767b2-a8de-4b37-a96b-e829ab251b46 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.804586] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 905.804586] env[63418]: value = "task-1245211" [ 905.804586] env[63418]: _type = "Task" [ 905.804586] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.812537] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245211, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.905544] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.905951] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5f894714-8dcb-4eb1-ad17-2ce3160c6edf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.914065] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 905.914065] env[63418]: value = "task-1245212" [ 905.914065] env[63418]: _type = "Task" [ 905.914065] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.923249] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245212, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.142338] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245209, 'name': Destroy_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.190942] env[63418]: DEBUG nova.network.neutron [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.275155] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.337s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.275390] env[63418]: INFO nova.compute.manager [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Migrating [ 906.275645] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.275796] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.277101] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245210, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.277558] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.999s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.277781] env[63418]: DEBUG nova.objects.instance [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lazy-loading 'resources' on Instance uuid 5ab04ddd-4cd6-4330-a9d7-081290c27586 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.318922] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245211, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.410825] env[63418]: DEBUG nova.network.neutron [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.426775] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245212, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.639423] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245209, 'name': Destroy_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.773685] env[63418]: DEBUG oslo_vmware.api [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245210, 'name': PowerOnVM_Task, 'duration_secs': 0.935112} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.774019] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.774261] env[63418]: INFO nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Took 8.75 seconds to spawn the instance on the hypervisor. [ 906.774456] env[63418]: DEBUG nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 906.775285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815a5d00-7416-4e10-92d8-c857603e4bd9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.777932] env[63418]: INFO nova.compute.rpcapi [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 906.778449] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.820428] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245211, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.913669] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.914230] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Instance network_info: |[{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 906.914784] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:f0:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b1f9284-9bb8-49b0-80f1-c2154e6ba534', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.924721] env[63418]: DEBUG oslo.service.loopingcall [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.927597] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.931821] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68fbf0a7-24bc-4ede-a29f-c5f921a986c2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.960459] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245212, 'name': CreateSnapshot_Task, 'duration_secs': 0.996126} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.960459] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.960459] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.960459] env[63418]: value = "task-1245213" [ 906.960459] env[63418]: _type = "Task" [ 906.960459] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.960459] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac25b4c-c677-4ebe-bfe8-f17d785d09bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.979044] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245213, 'name': CreateVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.041240] env[63418]: DEBUG nova.compute.manager [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 907.041445] env[63418]: DEBUG nova.compute.manager [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing instance network info cache due to event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 907.041660] env[63418]: DEBUG oslo_concurrency.lockutils [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.041810] env[63418]: DEBUG oslo_concurrency.lockutils [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.042035] env[63418]: DEBUG nova.network.neutron [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.046069] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341b6c15-76b5-40ae-b2c6-1560408830a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.054496] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0497028-d598-46e9-874f-e8d9ab9f7cd6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.091701] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00667d3b-8972-443d-8f7d-26184d9b4b4b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.099955] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186eea9d-4d63-40a1-99bf-4fe6e2b4c5f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.118124] env[63418]: DEBUG nova.compute.provider_tree [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.139788] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245209, 'name': Destroy_Task, 'duration_secs': 1.362974} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.140059] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroyed the VM [ 907.140392] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 907.140626] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2f4bea2a-5f39-4beb-a2ad-76da1b19a95f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.150156] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 907.150156] env[63418]: value = "task-1245214" [ 907.150156] env[63418]: _type = "Task" [ 907.150156] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.157052] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245214, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.302312] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.302548] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.302736] env[63418]: DEBUG nova.network.neutron [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.314802] env[63418]: INFO nova.compute.manager [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Took 28.70 seconds to build instance. [ 907.322597] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245211, 'name': CreateSnapshot_Task, 'duration_secs': 1.108369} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.322864] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 907.323692] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e106d83a-838b-4857-9b2a-02a7b01730bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.474725] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245213, 'name': CreateVM_Task, 'duration_secs': 0.34875} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.474839] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 907.476551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.476551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.476900] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.477333] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38f6bf02-b024-4718-947a-4455b3052bf7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.490601] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 907.492202] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ff26c6c4-ad6e-4703-84a5-f3cb54350324 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.495363] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 907.495363] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253006b-39c1-fc79-50af-5c62cf25375c" [ 907.495363] env[63418]: _type = "Task" [ 907.495363] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.501571] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 907.501571] env[63418]: value = "task-1245215" [ 907.501571] env[63418]: _type = "Task" [ 907.501571] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.509446] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253006b-39c1-fc79-50af-5c62cf25375c, 'name': SearchDatastore_Task, 'duration_secs': 0.01149} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.510148] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.510414] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.510678] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.510891] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.511191] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.511403] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4fde6cd-5fdd-4c64-ba7f-681a5da19b02 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.522346] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245215, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.532367] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.532646] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.533509] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ef37c4b-7977-492e-8c3a-8ebcc897b732 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.539025] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 907.539025] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520de855-9693-8060-0c44-6b3205bbb1e8" [ 907.539025] env[63418]: _type = "Task" [ 907.539025] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.549817] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520de855-9693-8060-0c44-6b3205bbb1e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.623062] env[63418]: DEBUG nova.scheduler.client.report [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 907.664353] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245214, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.729719] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.729999] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.817226] env[63418]: DEBUG oslo_concurrency.lockutils [None req-38f59588-92e1-422f-a1b3-d9505504cac0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.210s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.844723] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 907.848251] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e15a3b84-417a-4fbd-bd9e-19fdc39b89dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.856864] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 907.856864] env[63418]: value = "task-1245216" [ 907.856864] env[63418]: _type = "Task" [ 907.856864] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.865491] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245216, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.013448] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245215, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.050650] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520de855-9693-8060-0c44-6b3205bbb1e8, 'name': SearchDatastore_Task, 'duration_secs': 0.01132} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.051998] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9c8ae05-0e7c-48c3-bdd3-3cefa2f492d9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.059851] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 908.059851] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52309f35-060d-0373-924a-6f358889f42b" [ 908.059851] env[63418]: _type = "Task" [ 908.059851] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.074021] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52309f35-060d-0373-924a-6f358889f42b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.131392] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.133707] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.801s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.135747] env[63418]: INFO nova.compute.claims [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.159258] env[63418]: DEBUG oslo_vmware.api [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245214, 'name': RemoveSnapshot_Task, 'duration_secs': 0.675992} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.159592] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 908.159875] env[63418]: INFO nova.compute.manager [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 14.43 seconds to snapshot the instance on the hypervisor. [ 908.173109] env[63418]: INFO nova.scheduler.client.report [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleted allocations for instance 5ab04ddd-4cd6-4330-a9d7-081290c27586 [ 908.233940] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 908.344718] env[63418]: DEBUG nova.network.neutron [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updated VIF entry in instance network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.345228] env[63418]: DEBUG nova.network.neutron [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.372268] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245216, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.493427] env[63418]: DEBUG nova.network.neutron [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance_info_cache with network_info: [{"id": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "address": "fa:16:3e:72:aa:e2", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ff1dd4-2e", "ovs_interfaceid": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.516155] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245215, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.576442] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52309f35-060d-0373-924a-6f358889f42b, 'name': SearchDatastore_Task, 'duration_secs': 0.010948} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.576442] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.576659] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5/039f4cf6-137b-4356-a1b8-dd47f8f0d6f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.576738] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9aa970f9-dacf-4693-828d-d2461e29b64c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.584475] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 908.584475] env[63418]: value = "task-1245217" [ 908.584475] env[63418]: _type = "Task" [ 908.584475] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.594488] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.682138] env[63418]: DEBUG oslo_concurrency.lockutils [None req-44651dca-244f-4f3e-bada-795977862ac2 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "5ab04ddd-4cd6-4330-a9d7-081290c27586" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.887s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.720308] env[63418]: DEBUG nova.compute.manager [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Found 3 images (rotation: 2) {{(pid=63418) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 908.720521] env[63418]: DEBUG nova.compute.manager [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Rotating out 1 backups {{(pid=63418) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4905}} [ 908.720689] env[63418]: DEBUG nova.compute.manager [None req-3541e979-cf02-4f00-a61a-55ad6281ce3e tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleting image 218c0554-98fc-478a-bef5-e1a4088a8355 {{(pid=63418) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4910}} [ 908.743932] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "054bdcde-55a6-4f86-a230-92b2115e47b0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.745106] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.745486] env[63418]: DEBUG nova.compute.manager [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.749145] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958165a5-8605-45d3-9090-e12925815b4c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.766093] env[63418]: DEBUG nova.compute.manager [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 908.766822] env[63418]: DEBUG nova.objects.instance [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'flavor' on Instance uuid 054bdcde-55a6-4f86-a230-92b2115e47b0 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.782730] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.852141] env[63418]: DEBUG oslo_concurrency.lockutils [req-73cc9192-8331-4467-a649-c805e9a6fb83 req-20f4916e-9ce7-4f9a-af9d-c018f51588ce service nova] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.870982] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245216, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.996982] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.016469] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245215, 'name': CloneVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.094402] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245217, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.358884] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46cc8aeb-2f17-4206-8023-da63723e0c94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.381710] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984ac703-0b81-4fc6-ac07-ed79c83ed0a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.388426] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245216, 'name': CloneVM_Task, 'duration_secs': 1.407669} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.388867] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Created linked-clone VM from snapshot [ 909.390949] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4812e4-9be0-462d-83d2-92af8d62f3ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.448098] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6994dbd6-f368-44ac-8c38-f1825c009476 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.456817] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Uploading image 3e5a3485-02fb-4c87-bf16-417423b1a071 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 909.465908] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ba56a4-2446-44a6-b463-2d3cb915ed23 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.489859] env[63418]: DEBUG nova.compute.provider_tree [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.494497] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 909.494497] env[63418]: value = "vm-268500" [ 909.494497] env[63418]: _type = "VirtualMachine" [ 909.494497] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 909.494929] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5dbf2006-794d-4baf-a32a-bb9084dccbd2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.509027] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease: (returnval){ [ 909.509027] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5293b19a-ea98-9ce4-a109-3da9b1713559" [ 909.509027] env[63418]: _type = "HttpNfcLease" [ 909.509027] env[63418]: } obtained for exporting VM: (result){ [ 909.509027] env[63418]: value = "vm-268500" [ 909.509027] env[63418]: _type = "VirtualMachine" [ 909.509027] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 909.509027] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the lease: (returnval){ [ 909.509027] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5293b19a-ea98-9ce4-a109-3da9b1713559" [ 909.509027] env[63418]: _type = "HttpNfcLease" [ 909.509027] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 909.528017] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245215, 'name': CloneVM_Task, 'duration_secs': 1.731139} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.530254] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Created linked-clone VM from snapshot [ 909.532529] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.532529] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5293b19a-ea98-9ce4-a109-3da9b1713559" [ 909.532529] env[63418]: _type = "HttpNfcLease" [ 909.532529] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 909.532529] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed2cc78-ef84-4c9c-8ed8-2bd7aa2a8dc3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.536352] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 909.536352] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5293b19a-ea98-9ce4-a109-3da9b1713559" [ 909.536352] env[63418]: _type = "HttpNfcLease" [ 909.536352] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 909.538040] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a00b80a-0cdd-4055-ad91-5c94f5252e66 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.548530] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Uploading image 69a22df6-0b8e-4d81-ab99-3f97aaaed86e {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 909.554626] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52caa650-5448-2441-1c1c-31e343816e13/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 909.555032] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52caa650-5448-2441-1c1c-31e343816e13/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 909.631336] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519184} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.631609] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5/039f4cf6-137b-4356-a1b8-dd47f8f0d6f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.631827] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.632100] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a3e26f5-f98d-47e2-9b91-513eb5b9e7e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.636089] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 909.636089] env[63418]: value = "vm-268499" [ 909.636089] env[63418]: _type = "VirtualMachine" [ 909.636089] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 909.636356] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-301e655c-06b9-4e2a-88fa-fca192e09322 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.643516] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lease: (returnval){ [ 909.643516] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e0346-f7b1-65d5-493b-1c89947c2d4a" [ 909.643516] env[63418]: _type = "HttpNfcLease" [ 909.643516] env[63418]: } obtained for exporting VM: (result){ [ 909.643516] env[63418]: value = "vm-268499" [ 909.643516] env[63418]: _type = "VirtualMachine" [ 909.643516] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 909.643865] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the lease: (returnval){ [ 909.643865] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e0346-f7b1-65d5-493b-1c89947c2d4a" [ 909.643865] env[63418]: _type = "HttpNfcLease" [ 909.643865] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 909.645226] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 909.645226] env[63418]: value = "task-1245219" [ 909.645226] env[63418]: _type = "Task" [ 909.645226] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.659142] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.659142] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e0346-f7b1-65d5-493b-1c89947c2d4a" [ 909.659142] env[63418]: _type = "HttpNfcLease" [ 909.659142] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 909.659540] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245219, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.665283] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-43d7c4e1-e23f-4457-808a-4a5bdfd69089 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.774641] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.775043] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af6e4773-da15-4229-9468-a5c201eeb695 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.782378] env[63418]: DEBUG oslo_vmware.api [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 909.782378] env[63418]: value = "task-1245221" [ 909.782378] env[63418]: _type = "Task" [ 909.782378] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.791212] env[63418]: DEBUG oslo_vmware.api [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245221, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.999166] env[63418]: DEBUG nova.scheduler.client.report [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 910.161740] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.161740] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e0346-f7b1-65d5-493b-1c89947c2d4a" [ 910.161740] env[63418]: _type = "HttpNfcLease" [ 910.161740] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 910.164812] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 910.164812] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523e0346-f7b1-65d5-493b-1c89947c2d4a" [ 910.164812] env[63418]: _type = "HttpNfcLease" [ 910.164812] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 910.165363] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245219, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071237} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.166169] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e698e63-c15a-49dc-97cb-4e9b72136f7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.168930] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.169836] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d03649b-1948-420c-99c5-175abff8b1bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.179328] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b403c-d763-d8e9-263f-e049b1926d6b/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 910.179673] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b403c-d763-d8e9-263f-e049b1926d6b/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 910.200316] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5/039f4cf6-137b-4356-a1b8-dd47f8f0d6f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.201702] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19a4bfe6-e12d-44ae-82e6-249c05ae77be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.276888] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 910.276888] env[63418]: value = "task-1245222" [ 910.276888] env[63418]: _type = "Task" [ 910.276888] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.287248] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245222, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.296535] env[63418]: DEBUG oslo_vmware.api [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245221, 'name': PowerOffVM_Task, 'duration_secs': 0.291415} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.298119] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.298119] env[63418]: DEBUG nova.compute.manager [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 910.298581] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfda34a-50be-4ec5-af1f-b5f17d552d6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.341217] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3769ef33-9570-4d79-986e-515e0e705330 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.507493] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.508921] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 910.513658] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.731s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.517098] env[63418]: INFO nova.compute.claims [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.534292] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ec56fc-1226-41d7-b335-c37ddf32f09c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.564023] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 0 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 910.642940] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "4838d2b2-a187-48ae-a3b1-8ca24b961359" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.643087] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.787728] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245222, 'name': ReconfigVM_Task, 'duration_secs': 0.341259} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.788103] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5/039f4cf6-137b-4356-a1b8-dd47f8f0d6f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.788778] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-848c1ea9-e9de-42ef-8f60-c334c3c28c4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.796788] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 910.796788] env[63418]: value = "task-1245223" [ 910.796788] env[63418]: _type = "Task" [ 910.796788] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.807446] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245223, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.811467] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c03b4e29-a01c-472c-891f-83fc58c96186 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.067s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.023519] env[63418]: DEBUG nova.compute.utils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.026727] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 911.027588] env[63418]: DEBUG nova.network.neutron [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.072307] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.074624] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe6cac71-31ff-4ee1-a97c-72dbe04658ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.083964] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 911.083964] env[63418]: value = "task-1245224" [ 911.083964] env[63418]: _type = "Task" [ 911.083964] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.097465] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.104332] env[63418]: DEBUG nova.policy [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.146633] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 911.310822] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245223, 'name': Rename_Task, 'duration_secs': 0.175992} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.312110] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.312553] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11e99351-529a-4138-8c05-061a203d276b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.320120] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 911.320120] env[63418]: value = "task-1245225" [ 911.320120] env[63418]: _type = "Task" [ 911.320120] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.335420] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.532174] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 911.602975] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245224, 'name': PowerOffVM_Task, 'duration_secs': 0.221775} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.603319] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.604764] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 17 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 911.651673] env[63418]: DEBUG nova.network.neutron [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Successfully created port: 4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.680383] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.759501] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "054bdcde-55a6-4f86-a230-92b2115e47b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.759797] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.760068] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "054bdcde-55a6-4f86-a230-92b2115e47b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.760319] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.760556] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.765781] env[63418]: INFO nova.compute.manager [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Terminating instance [ 911.834639] env[63418]: DEBUG oslo_vmware.api [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245225, 'name': PowerOnVM_Task, 'duration_secs': 0.511443} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.835356] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.835722] env[63418]: INFO nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Took 7.72 seconds to spawn the instance on the hypervisor. [ 911.836250] env[63418]: DEBUG nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 911.842802] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc07c344-b498-46a4-82af-74a9d3736926 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.868298] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76599b9e-c2ae-4e06-981e-c99782136f16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.877112] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580e9dcc-8cb5-4d8f-897a-5c0458607f9e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.914043] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8384840c-fa27-4e89-81b2-0b2fe93451a1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.925865] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e30ac98-c444-48c1-9e21-5fceca7be7c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.939701] env[63418]: DEBUG nova.compute.provider_tree [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.117494] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.117875] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.118172] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.118390] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.119031] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.119197] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.119490] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.120155] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.120230] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.120441] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.120808] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.127492] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.127918] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.128177] env[63418]: DEBUG nova.compute.manager [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 912.128955] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f63eee8-0802-4e01-9d0b-629a96af8bbd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.141148] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd07e1b-f775-405c-833b-873aa9127841 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.148422] env[63418]: DEBUG nova.compute.manager [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 912.149091] env[63418]: DEBUG nova.objects.instance [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'flavor' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.152411] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 912.152411] env[63418]: value = "task-1245226" [ 912.152411] env[63418]: _type = "Task" [ 912.152411] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.162229] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245226, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.270497] env[63418]: DEBUG nova.compute.manager [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 912.270497] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.270497] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127a8646-fa10-4409-b19f-6c77c136b276 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.280896] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.281545] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07f00234-4d05-4f48-8347-a385cf59722a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.371301] env[63418]: INFO nova.compute.manager [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Took 18.20 seconds to build instance. [ 912.374338] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.374640] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.374974] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleting the datastore file [datastore2] 054bdcde-55a6-4f86-a230-92b2115e47b0 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.375707] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6df690f-5153-439d-a283-763f1fd2154e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.382830] env[63418]: DEBUG oslo_vmware.api [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 912.382830] env[63418]: value = "task-1245228" [ 912.382830] env[63418]: _type = "Task" [ 912.382830] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.395514] env[63418]: DEBUG oslo_vmware.api [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245228, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.443788] env[63418]: DEBUG nova.scheduler.client.report [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 912.547776] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 912.664318] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245226, 'name': ReconfigVM_Task, 'duration_secs': 0.224471} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.664883] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 33 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 912.880868] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f5f19700-8934-4043-a88f-89f6967c0857 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.711s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.896718] env[63418]: DEBUG oslo_vmware.api [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245228, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151337} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.896978] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.898147] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 912.898147] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 912.898147] env[63418]: INFO nova.compute.manager [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Took 0.63 seconds to destroy the instance on the hypervisor. [ 912.898147] env[63418]: DEBUG oslo.service.loopingcall [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.898147] env[63418]: DEBUG nova.compute.manager [-] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 912.898147] env[63418]: DEBUG nova.network.neutron [-] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 912.950373] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.951148] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 912.957309] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.277s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.959042] env[63418]: INFO nova.compute.claims [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.161916] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.162308] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c978ade-7ce1-4b43-88c0-6d8366d61a7c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.174617] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.175059] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.175147] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.175297] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.175496] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.175910] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.176178] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.176355] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.176627] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.176696] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.177133] env[63418]: DEBUG nova.virt.hardware [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.184412] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Reconfiguring VM instance instance-0000004f to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 913.184840] env[63418]: DEBUG oslo_vmware.api [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 913.184840] env[63418]: value = "task-1245229" [ 913.184840] env[63418]: _type = "Task" [ 913.184840] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.185122] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a822c01-29d8-449a-82bb-bc34ffc6a943 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.211405] env[63418]: DEBUG oslo_vmware.api [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.212945] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 913.212945] env[63418]: value = "task-1245230" [ 913.212945] env[63418]: _type = "Task" [ 913.212945] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.222843] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245230, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.465298] env[63418]: DEBUG nova.compute.utils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.473722] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 913.473937] env[63418]: DEBUG nova.network.neutron [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.562443] env[63418]: DEBUG nova.policy [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b0f6e35d8494335b25867ab54fe1a3d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e074b295ca4d4e50bde88f6d37a7f98f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.707692] env[63418]: DEBUG nova.compute.manager [req-fd1ef462-8b02-486b-9f15-1c5ef554ae83 req-9d6c202e-a2ae-47b3-8ce4-ca0dde723039 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Received event network-vif-deleted-de942ff3-dc1e-4603-9794-16ca7223fcb2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 913.707903] env[63418]: INFO nova.compute.manager [req-fd1ef462-8b02-486b-9f15-1c5ef554ae83 req-9d6c202e-a2ae-47b3-8ce4-ca0dde723039 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Neutron deleted interface de942ff3-dc1e-4603-9794-16ca7223fcb2; detaching it from the instance and deleting it from the info cache [ 913.708144] env[63418]: DEBUG nova.network.neutron [req-fd1ef462-8b02-486b-9f15-1c5ef554ae83 req-9d6c202e-a2ae-47b3-8ce4-ca0dde723039 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.717213] env[63418]: DEBUG oslo_vmware.api [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245229, 'name': PowerOffVM_Task, 'duration_secs': 0.289666} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.726190] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.726388] env[63418]: DEBUG nova.compute.manager [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 913.728993] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf43e85-e527-435c-9039-bbc793909038 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.741505] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245230, 'name': ReconfigVM_Task, 'duration_secs': 0.241249} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.743038] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Reconfigured VM instance instance-0000004f to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 913.747429] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7855eecf-95da-450d-ae84-d06c261d17b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.772283] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5/1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.773562] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98a15b1a-126b-405a-9a01-583de7db9a07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.794727] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 913.794727] env[63418]: value = "task-1245231" [ 913.794727] env[63418]: _type = "Task" [ 913.794727] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.803793] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245231, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.806506] env[63418]: DEBUG nova.network.neutron [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Successfully updated port: 4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.887430] env[63418]: DEBUG nova.network.neutron [-] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.948855] env[63418]: DEBUG nova.network.neutron [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Successfully created port: ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.981640] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 914.213198] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddbb8967-ed78-4212-92a7-96aecc15af1d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.221388] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ac7baa-afca-4005-9f18-5c23b2eff515 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.233893] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcae6e4-9a68-4d3f-9454-287fdec7b362 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.242196] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f59c19-b9bd-4193-b437-1a1d8f759dcf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.259614] env[63418]: DEBUG nova.compute.manager [req-fd1ef462-8b02-486b-9f15-1c5ef554ae83 req-9d6c202e-a2ae-47b3-8ce4-ca0dde723039 service nova] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Detach interface failed, port_id=de942ff3-dc1e-4603-9794-16ca7223fcb2, reason: Instance 054bdcde-55a6-4f86-a230-92b2115e47b0 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 914.260599] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6aef1d7f-c056-4f20-bd74-d24aacde482b tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.132s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.290011] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfebf2a2-4da0-4de4-9fe3-534e3afaf11c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.302497] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4382a823-a7e3-4185-a3c4-7e4d30b3019c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.312829] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-14832e8e-3b4a-462e-84bb-d5a1e541d329" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.313036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-14832e8e-3b4a-462e-84bb-d5a1e541d329" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.313240] env[63418]: DEBUG nova.network.neutron [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.314664] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245231, 'name': ReconfigVM_Task, 'duration_secs': 0.462205} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.315542] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5/1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.315838] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 50 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 914.328393] env[63418]: DEBUG nova.compute.provider_tree [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.391021] env[63418]: INFO nova.compute.manager [-] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Took 1.49 seconds to deallocate network for instance. [ 914.833403] env[63418]: DEBUG nova.scheduler.client.report [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 914.838582] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc7ff38-f496-4059-84ea-17da90fb9b6e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.861260] env[63418]: DEBUG nova.network.neutron [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.865299] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a5b8a6-e6ba-4d97-be29-e3e09eb4079d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.888989] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 67 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 914.898994] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.996093] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 915.039538] env[63418]: DEBUG nova.network.neutron [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Updating instance_info_cache with network_info: [{"id": "4fdaa32d-3674-4689-8a75-07aa2e067de4", "address": "fa:16:3e:4a:b3:ef", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fdaa32d-36", "ovs_interfaceid": "4fdaa32d-3674-4689-8a75-07aa2e067de4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.344573] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.345145] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 915.348270] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.449s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.348487] env[63418]: DEBUG nova.objects.instance [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'resources' on Instance uuid 054bdcde-55a6-4f86-a230-92b2115e47b0 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.461768] env[63418]: DEBUG nova.network.neutron [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Port e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 binding to destination host cpu-1 is already ACTIVE {{(pid=63418) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 915.542227] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-14832e8e-3b4a-462e-84bb-d5a1e541d329" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.542831] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Instance network_info: |[{"id": "4fdaa32d-3674-4689-8a75-07aa2e067de4", "address": "fa:16:3e:4a:b3:ef", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fdaa32d-36", "ovs_interfaceid": "4fdaa32d-3674-4689-8a75-07aa2e067de4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 915.543349] env[63418]: DEBUG nova.network.neutron [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Successfully updated port: ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.852031] env[63418]: DEBUG nova.compute.utils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.857067] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 915.857473] env[63418]: DEBUG nova.network.neutron [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.898495] env[63418]: DEBUG nova.policy [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2dcdc53fd7cc4555a4eecb6bf7564b55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4759c1d6199344c49063d35e422d6324', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.046036] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f54095d-58f6-46f6-8479-080b8d5990e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.050624] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.051127] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.051417] env[63418]: DEBUG nova.network.neutron [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.059193] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59af74d1-dfdb-4f19-ba1d-cf1121ef6d06 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.097419] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae82a2bf-4cb7-4faf-8aa7-653121e8fc81 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.106441] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39d2cf2-5a6b-4f4d-a96e-b80c3228ec60 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.121279] env[63418]: DEBUG nova.compute.provider_tree [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.165804] env[63418]: DEBUG nova.network.neutron [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Successfully created port: fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.357522] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 916.487396] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.487644] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.487819] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.586529] env[63418]: DEBUG nova.network.neutron [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 916.626823] env[63418]: DEBUG nova.scheduler.client.report [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 916.714080] env[63418]: DEBUG nova.network.neutron [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updating instance_info_cache with network_info: [{"id": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "address": "fa:16:3e:36:3a:6f", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef27d89f-f8", "ovs_interfaceid": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.132154] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.159886] env[63418]: INFO nova.scheduler.client.report [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocations for instance 054bdcde-55a6-4f86-a230-92b2115e47b0 [ 917.217312] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.217677] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Instance network_info: |[{"id": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "address": "fa:16:3e:36:3a:6f", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef27d89f-f8", "ovs_interfaceid": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 917.368407] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 917.562162] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.562404] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.562596] env[63418]: DEBUG nova.network.neutron [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.668949] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8777bab0-37c8-476c-812f-17d6275b1e1b tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "054bdcde-55a6-4f86-a230-92b2115e47b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.909s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.711710] env[63418]: DEBUG nova.network.neutron [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Successfully updated port: fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.215477] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "refresh_cache-4838d2b2-a187-48ae-a3b1-8ca24b961359" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.215742] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "refresh_cache-4838d2b2-a187-48ae-a3b1-8ca24b961359" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.215880] env[63418]: DEBUG nova.network.neutron [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.364598] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.364876] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.365128] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.365392] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.365554] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.365709] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.365924] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.366099] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.366278] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.366456] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.366664] env[63418]: DEBUG nova.virt.hardware [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.369079] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b2c8c5-8eb0-40cc-b5f1-1dceeeae7126 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.381755] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.382075] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.382321] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.383298] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.383691] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.384419] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.384858] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.385218] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.385419] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.386363] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.386741] env[63418]: DEBUG nova.virt.hardware [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.388586] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2547ab70-8d2d-4d0e-a73a-43f73e2d646b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.394380] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e05175d-fee9-4b82-aafb-deaf53f99d49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.403623] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.403900] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.404083] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.404298] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.404529] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.404703] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.404915] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.405093] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.405267] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.405436] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.405609] env[63418]: DEBUG nova.virt.hardware [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.407482] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0559ca-864b-4e01-80d1-d880d7539aa4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.421256] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:b3:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4fdaa32d-3674-4689-8a75-07aa2e067de4', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.429094] env[63418]: DEBUG oslo.service.loopingcall [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.431357] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52caa650-5448-2441-1c1c-31e343816e13/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.433043] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195ac115-567e-4690-a546-11dc563039c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.437085] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.437863] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f918e1a2-baf0-464f-9c2d-ddfe921e78a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.444602] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff8ebadd-8081-4d1d-b741-7b8488a8073e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.460503] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749c479b-aa89-4ad6-9ab0-45ec6f097c14 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.473985] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:3a:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b36c5ae6-c344-4bd1-8239-29128e2bbfbf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef27d89f-f8f7-46cc-92e2-811aee3d2dea', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.481623] env[63418]: DEBUG oslo.service.loopingcall [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.481944] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52caa650-5448-2441-1c1c-31e343816e13/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.482116] env[63418]: ERROR oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52caa650-5448-2441-1c1c-31e343816e13/disk-0.vmdk due to incomplete transfer. [ 918.483779] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.483991] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1e49c703-e069-415f-b5b3-f27cef38de81 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.485969] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.485969] env[63418]: value = "task-1245232" [ 918.485969] env[63418]: _type = "Task" [ 918.485969] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.493764] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e40d6fa-862f-4cf6-a0e0-05478aae4dea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.509474] env[63418]: DEBUG nova.network.neutron [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance_info_cache with network_info: [{"id": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "address": "fa:16:3e:72:aa:e2", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ff1dd4-2e", "ovs_interfaceid": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.516671] env[63418]: DEBUG oslo_vmware.rw_handles [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52caa650-5448-2441-1c1c-31e343816e13/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 918.516874] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Uploaded image 3e5a3485-02fb-4c87-bf16-417423b1a071 to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 918.519347] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 918.523573] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cf2f39aa-dc23-4a1a-b1b5-b3a30a1f2cf8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.525132] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245232, 'name': CreateVM_Task} progress is 15%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.525358] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.525358] env[63418]: value = "task-1245233" [ 918.525358] env[63418]: _type = "Task" [ 918.525358] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.530854] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 918.530854] env[63418]: value = "task-1245234" [ 918.530854] env[63418]: _type = "Task" [ 918.530854] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.537870] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245233, 'name': CreateVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.545729] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245234, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.753253] env[63418]: DEBUG nova.network.neutron [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.847625] env[63418]: DEBUG nova.compute.manager [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 918.848402] env[63418]: DEBUG nova.compute.manager [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing instance network info cache due to event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 918.848402] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.848402] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.848402] env[63418]: DEBUG nova.network.neutron [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.914459] env[63418]: DEBUG nova.network.neutron [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Updating instance_info_cache with network_info: [{"id": "fc06d529-525c-4dbf-8050-32c6054ecc96", "address": "fa:16:3e:b5:18:2e", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc06d529-52", "ovs_interfaceid": "fc06d529-525c-4dbf-8050-32c6054ecc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.015065] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.025325] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245232, 'name': CreateVM_Task, 'duration_secs': 0.474753} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.026350] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.029841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.029841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.029841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.029841] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8910a25c-4b56-49d0-86ee-be04ae1a4f4b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.041117] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 919.041117] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b3e50-fb20-f353-50bc-cf7e2dd6cee6" [ 919.041117] env[63418]: _type = "Task" [ 919.041117] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.045015] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245233, 'name': CreateVM_Task, 'duration_secs': 0.359376} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.048780] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.052457] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.052742] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245234, 'name': Destroy_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.058560] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b3e50-fb20-f353-50bc-cf7e2dd6cee6, 'name': SearchDatastore_Task, 'duration_secs': 0.010011} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.058879] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.059130] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.059432] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.059592] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.059778] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.060121] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.060438] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.060690] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-233b3da1-a49f-47f2-b4d1-98fbe7a83e0f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.062644] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a0845f8-f38b-4a18-b2f5-a7940ac7c2d6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.068089] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 919.068089] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292086-3c02-3f44-4226-efa99857c596" [ 919.068089] env[63418]: _type = "Task" [ 919.068089] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.072182] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.072438] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.073559] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cca852a-c47d-4c76-b2dd-3f9c8fb2c570 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.079564] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292086-3c02-3f44-4226-efa99857c596, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.082639] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 919.082639] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526e0bde-1988-1fb4-48dd-2407d66ce133" [ 919.082639] env[63418]: _type = "Task" [ 919.082639] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.091220] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526e0bde-1988-1fb4-48dd-2407d66ce133, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.421039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "refresh_cache-4838d2b2-a187-48ae-a3b1-8ca24b961359" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.421373] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Instance network_info: |[{"id": "fc06d529-525c-4dbf-8050-32c6054ecc96", "address": "fa:16:3e:b5:18:2e", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc06d529-52", "ovs_interfaceid": "fc06d529-525c-4dbf-8050-32c6054ecc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 919.421828] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:18:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fc06d529-525c-4dbf-8050-32c6054ecc96', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.429991] env[63418]: DEBUG oslo.service.loopingcall [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.433756] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.437429] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d10d5ac8-ca68-435e-8e84-80ffd79e80d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.455374] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.455759] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.458415] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Received event network-vif-plugged-4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 919.459041] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Acquiring lock "14832e8e-3b4a-462e-84bb-d5a1e541d329-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.459041] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.459155] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.459365] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] No waiting events found dispatching network-vif-plugged-4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 919.459583] env[63418]: WARNING nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Received unexpected event network-vif-plugged-4fdaa32d-3674-4689-8a75-07aa2e067de4 for instance with vm_state building and task_state spawning. [ 919.459799] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Received event network-changed-4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 919.460114] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Refreshing instance network info cache due to event network-changed-4fdaa32d-3674-4689-8a75-07aa2e067de4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 919.460370] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Acquiring lock "refresh_cache-14832e8e-3b4a-462e-84bb-d5a1e541d329" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.460968] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Acquired lock "refresh_cache-14832e8e-3b4a-462e-84bb-d5a1e541d329" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.461218] env[63418]: DEBUG nova.network.neutron [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Refreshing network info cache for port 4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.470037] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.470037] env[63418]: value = "task-1245235" [ 919.470037] env[63418]: _type = "Task" [ 919.470037] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.478995] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245235, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.546552] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245234, 'name': Destroy_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.550906] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e81d33e-2ce9-44be-88e6-eb8b82504616 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.576935] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a574ac-a337-4877-bd0c-d7d99132d8a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.585639] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292086-3c02-3f44-4226-efa99857c596, 'name': SearchDatastore_Task, 'duration_secs': 0.00889} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.591661] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.591866] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.592401] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.592546] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 83 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 919.605857] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526e0bde-1988-1fb4-48dd-2407d66ce133, 'name': SearchDatastore_Task, 'duration_secs': 0.008975} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.607953] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae00ae46-4c67-4f98-a7ca-ef84fa13b640 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.612742] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 919.612742] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c4ab3c-25af-be02-715c-75f1b8ed8be5" [ 919.612742] env[63418]: _type = "Task" [ 919.612742] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.624028] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c4ab3c-25af-be02-715c-75f1b8ed8be5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.684461] env[63418]: DEBUG nova.network.neutron [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updated VIF entry in instance network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.684934] env[63418]: DEBUG nova.network.neutron [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.715078] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b403c-d763-d8e9-263f-e049b1926d6b/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 919.715839] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.716079] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.718227] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6fa778-1e1c-4b3b-b7f4-9d9a00d49b81 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.725457] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b403c-d763-d8e9-263f-e049b1926d6b/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 919.725647] env[63418]: ERROR oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b403c-d763-d8e9-263f-e049b1926d6b/disk-0.vmdk due to incomplete transfer. [ 919.726888] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d74617dd-f2fa-49a8-a44a-dfb5cc85bff6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.734196] env[63418]: DEBUG oslo_vmware.rw_handles [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b403c-d763-d8e9-263f-e049b1926d6b/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 919.734486] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Uploaded image 69a22df6-0b8e-4d81-ab99-3f97aaaed86e to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 919.736074] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 919.736730] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-70cc2a7f-4bbc-475d-bacd-4ad82f9d24f9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.742127] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 919.742127] env[63418]: value = "task-1245236" [ 919.742127] env[63418]: _type = "Task" [ 919.742127] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.750236] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245236, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.969323] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 919.986413] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245235, 'name': CreateVM_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.049809] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245234, 'name': Destroy_Task, 'duration_secs': 1.454579} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.050098] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Destroyed the VM [ 920.050380] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.050604] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0f9c4c2c-d2c1-418e-bf4a-46d4993c8315 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.059045] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 920.059045] env[63418]: value = "task-1245237" [ 920.059045] env[63418]: _type = "Task" [ 920.059045] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.067520] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245237, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.103217] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.103664] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb1942f2-2bc5-4fd0-8b37-703f94dbfc4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.114209] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 920.114209] env[63418]: value = "task-1245238" [ 920.114209] env[63418]: _type = "Task" [ 920.114209] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.128029] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c4ab3c-25af-be02-715c-75f1b8ed8be5, 'name': SearchDatastore_Task, 'duration_secs': 0.025693} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.132507] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.132978] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 14832e8e-3b4a-462e-84bb-d5a1e541d329/14832e8e-3b4a-462e-84bb-d5a1e541d329.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.133274] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245238, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.133734] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.134165] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.134449] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c90af80-f055-4309-ab3a-96f4d086618d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.139466] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f72a96ce-bfb7-4ebf-a686-34330b483ae9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.145770] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 920.145770] env[63418]: value = "task-1245239" [ 920.145770] env[63418]: _type = "Task" [ 920.145770] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.151361] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.151563] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.153210] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e4e8138-2972-47f4-8875-45af179303f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.160369] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.163786] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 920.163786] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a92e3f-bbeb-991b-e897-42759b5d75ba" [ 920.163786] env[63418]: _type = "Task" [ 920.163786] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.173054] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a92e3f-bbeb-991b-e897-42759b5d75ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.188722] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bbe953e-0889-406b-ae12-100f02148607 req-09575f23-1f5f-4a1c-a60a-e79502ac1efa service nova] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.218392] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 920.255972] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245236, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.260688] env[63418]: DEBUG nova.network.neutron [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Updated VIF entry in instance network info cache for port 4fdaa32d-3674-4689-8a75-07aa2e067de4. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.261153] env[63418]: DEBUG nova.network.neutron [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Updating instance_info_cache with network_info: [{"id": "4fdaa32d-3674-4689-8a75-07aa2e067de4", "address": "fa:16:3e:4a:b3:ef", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fdaa32d-36", "ovs_interfaceid": "4fdaa32d-3674-4689-8a75-07aa2e067de4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.487361] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245235, 'name': CreateVM_Task, 'duration_secs': 0.81795} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.490512] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.492130] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.492130] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.492611] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.493543] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a75d80-97f5-408a-bfae-3de182235b0e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.500291] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 920.500291] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52487bc8-2115-553d-3b5e-9d15577e48a8" [ 920.500291] env[63418]: _type = "Task" [ 920.500291] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.506033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.506442] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.510284] env[63418]: INFO nova.compute.claims [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.514471] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52487bc8-2115-553d-3b5e-9d15577e48a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.570843] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245237, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.627374] env[63418]: DEBUG oslo_vmware.api [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245238, 'name': PowerOnVM_Task, 'duration_secs': 0.494581} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.627374] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.627545] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f6355750-c0c9-4597-8df8-c6f11786bbf1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance '1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5' progress to 100 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 920.656228] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245239, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488514} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.656494] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 14832e8e-3b4a-462e-84bb-d5a1e541d329/14832e8e-3b4a-462e-84bb-d5a1e541d329.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.656745] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.657013] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9b703c5-9f65-4f9c-b964-2bca574e0b04 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.662617] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 920.662617] env[63418]: value = "task-1245240" [ 920.662617] env[63418]: _type = "Task" [ 920.662617] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.675033] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245240, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.677550] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a92e3f-bbeb-991b-e897-42759b5d75ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009593} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.679596] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b5184e5-bb85-4dc3-b05c-56bb00adf682 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.685831] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 920.685831] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abecd0-830d-fd59-f9c7-d0c4a7b9d369" [ 920.685831] env[63418]: _type = "Task" [ 920.685831] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.693770] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abecd0-830d-fd59-f9c7-d0c4a7b9d369, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.712044] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "26147513-3b25-4cc1-991a-34a724f73711" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.712363] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "26147513-3b25-4cc1-991a-34a724f73711" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.719646] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "23af1d2e-f8ec-4c09-a070-9719b593f83a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.719912] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.741432] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.752305] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245236, 'name': Destroy_Task, 'duration_secs': 0.695754} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.752555] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Destroyed the VM [ 920.752790] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.753043] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-35679673-86fb-4f71-ad78-0380bce25f1d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.759273] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 920.759273] env[63418]: value = "task-1245241" [ 920.759273] env[63418]: _type = "Task" [ 920.759273] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.767491] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Releasing lock "refresh_cache-14832e8e-3b4a-462e-84bb-d5a1e541d329" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.767770] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Received event network-vif-plugged-ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 920.767992] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Acquiring lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.768218] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.768411] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.768673] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] No waiting events found dispatching network-vif-plugged-ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 920.768919] env[63418]: WARNING nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Received unexpected event network-vif-plugged-ef27d89f-f8f7-46cc-92e2-811aee3d2dea for instance with vm_state building and task_state spawning. [ 920.769111] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Received event network-changed-ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 920.769341] env[63418]: DEBUG nova.compute.manager [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Refreshing instance network info cache due to event network-changed-ef27d89f-f8f7-46cc-92e2-811aee3d2dea. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 920.769512] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Acquiring lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.769661] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Acquired lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.769826] env[63418]: DEBUG nova.network.neutron [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Refreshing network info cache for port ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.771017] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245241, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.896439] env[63418]: DEBUG nova.compute.manager [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Stashing vm_state: stopped {{(pid=63418) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 921.011600] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52487bc8-2115-553d-3b5e-9d15577e48a8, 'name': SearchDatastore_Task, 'duration_secs': 0.058351} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.011973] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.012283] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.012504] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.067850] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245237, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.171571] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066805} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.171779] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.172538] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92656404-c152-42b6-85f0-ac7fd16cdcb3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.194143] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 14832e8e-3b4a-462e-84bb-d5a1e541d329/14832e8e-3b4a-462e-84bb-d5a1e541d329.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.194664] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04994919-00fb-4acc-b16d-7022e6df667e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.216896] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 921.219464] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abecd0-830d-fd59-f9c7-d0c4a7b9d369, 'name': SearchDatastore_Task, 'duration_secs': 0.009416} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.221229] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.221511] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/24a97d0a-252b-4bbd-9dfc-57767ef8014d.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.221848] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 921.221848] env[63418]: value = "task-1245242" [ 921.221848] env[63418]: _type = "Task" [ 921.221848] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.222084] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.222298] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.222536] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2b870b8-5a91-43fd-809a-1296962b1b2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.224542] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 921.227261] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c134c6ea-3eda-4754-86f7-60086c9b0862 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.238699] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245242, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.242392] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 921.242392] env[63418]: value = "task-1245243" [ 921.242392] env[63418]: _type = "Task" [ 921.242392] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.242392] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.242392] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.242392] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78471000-799f-4655-a754-3a704ba9de4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.250396] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 921.250396] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5254f351-f3f2-72ed-1ece-fef33e44c417" [ 921.250396] env[63418]: _type = "Task" [ 921.250396] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.254034] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245243, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.261618] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5254f351-f3f2-72ed-1ece-fef33e44c417, 'name': SearchDatastore_Task} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.266658] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-915042a9-ced5-46d1-b44c-c02df04d08e7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.277423] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245241, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.278072] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 921.278072] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521250f6-c278-e602-8a00-bcbd9d00a7fe" [ 921.278072] env[63418]: _type = "Task" [ 921.278072] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.286402] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521250f6-c278-e602-8a00-bcbd9d00a7fe, 'name': SearchDatastore_Task, 'duration_secs': 0.008646} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.286987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.287272] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 4838d2b2-a187-48ae-a3b1-8ca24b961359/4838d2b2-a187-48ae-a3b1-8ca24b961359.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.287580] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-192df89a-3e12-4c34-b5c6-d5b7fb7548fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.293693] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 921.293693] env[63418]: value = "task-1245244" [ 921.293693] env[63418]: _type = "Task" [ 921.293693] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.297061] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.297329] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.303830] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245244, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.415760] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.539835] env[63418]: DEBUG nova.network.neutron [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updated VIF entry in instance network info cache for port ef27d89f-f8f7-46cc-92e2-811aee3d2dea. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.540218] env[63418]: DEBUG nova.network.neutron [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updating instance_info_cache with network_info: [{"id": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "address": "fa:16:3e:36:3a:6f", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef27d89f-f8", "ovs_interfaceid": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.577465] env[63418]: DEBUG oslo_vmware.api [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245237, 'name': RemoveSnapshot_Task, 'duration_secs': 1.321965} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.577755] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 921.578103] env[63418]: INFO nova.compute.manager [None req-567a8b46-02b7-4876-b4f8-c863f75aeda2 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Took 16.32 seconds to snapshot the instance on the hypervisor. [ 921.633324] env[63418]: DEBUG nova.compute.manager [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Received event network-vif-plugged-fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 921.634028] env[63418]: DEBUG oslo_concurrency.lockutils [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] Acquiring lock "4838d2b2-a187-48ae-a3b1-8ca24b961359-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.634028] env[63418]: DEBUG oslo_concurrency.lockutils [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.634608] env[63418]: DEBUG oslo_concurrency.lockutils [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.634608] env[63418]: DEBUG nova.compute.manager [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] No waiting events found dispatching network-vif-plugged-fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 921.634727] env[63418]: WARNING nova.compute.manager [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Received unexpected event network-vif-plugged-fc06d529-525c-4dbf-8050-32c6054ecc96 for instance with vm_state building and task_state spawning. [ 921.634957] env[63418]: DEBUG nova.compute.manager [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Received event network-changed-fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 921.635249] env[63418]: DEBUG nova.compute.manager [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Refreshing instance network info cache due to event network-changed-fc06d529-525c-4dbf-8050-32c6054ecc96. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 921.636076] env[63418]: DEBUG oslo_concurrency.lockutils [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] Acquiring lock "refresh_cache-4838d2b2-a187-48ae-a3b1-8ca24b961359" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.636822] env[63418]: DEBUG oslo_concurrency.lockutils [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] Acquired lock "refresh_cache-4838d2b2-a187-48ae-a3b1-8ca24b961359" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.637119] env[63418]: DEBUG nova.network.neutron [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Refreshing network info cache for port fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.748247] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245242, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.761617] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.774667] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245243, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.781052] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.788968] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245241, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.804908] env[63418]: INFO nova.compute.manager [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Detaching volume ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967 [ 921.817357] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245244, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.860535] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f898a233-8d15-4b34-a317-330b5d4c6b54 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.869433] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a127cf-8f87-47d4-b24a-44e5b504cdb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.874515] env[63418]: INFO nova.virt.block_device [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Attempting to driver detach volume ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967 from mountpoint /dev/sdb [ 921.877454] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 921.877454] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268463', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'name': 'volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'd76a008c-9bd9-420b-873d-4f7d7f25b8ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'serial': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 921.877454] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a63f621-e5a3-4514-a52a-be832c08cd0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.928895] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507771c4-a1ee-4773-9bf3-c83da5f140f1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.932695] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625e6452-db49-46f3-a8c2-7c723bdc2e12 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.944256] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19f2145-4ed3-4593-b1cc-f442ae3aa16e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.948215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715f2d8f-866d-4015-8a47-d12072ab239c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.963865] env[63418]: DEBUG nova.compute.provider_tree [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.983990] env[63418]: DEBUG nova.scheduler.client.report [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 921.988223] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee423346-0a28-4455-8edc-bfad74cb59ad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.006243] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] The volume has not been displaced from its original location: [datastore1] volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967/volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 922.012306] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfiguring VM instance instance-00000031 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 922.013063] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe11dfe7-819f-44fb-a1e5-755f384c6bf0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.032699] env[63418]: DEBUG oslo_vmware.api [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 922.032699] env[63418]: value = "task-1245245" [ 922.032699] env[63418]: _type = "Task" [ 922.032699] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.040819] env[63418]: DEBUG oslo_vmware.api [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245245, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.043422] env[63418]: DEBUG oslo_concurrency.lockutils [req-7633e0f4-80e4-447e-9ec8-fc16e51e814d req-ad32f0e5-68ac-4918-968f-61741dd35780 service nova] Releasing lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.240709] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245242, 'name': ReconfigVM_Task, 'duration_secs': 0.568944} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.240997] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 14832e8e-3b4a-462e-84bb-d5a1e541d329/14832e8e-3b4a-462e-84bb-d5a1e541d329.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.241629] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aef01ed6-ebb0-4383-ab63-2bb41da184fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.247085] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 922.247085] env[63418]: value = "task-1245246" [ 922.247085] env[63418]: _type = "Task" [ 922.247085] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.258905] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245246, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.261806] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245243, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542444} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.262051] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/24a97d0a-252b-4bbd-9dfc-57767ef8014d.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.262441] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.262730] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3fb8fc39-8ffc-4fce-814d-681936bc111b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.272743] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245241, 'name': RemoveSnapshot_Task, 'duration_secs': 1.082078} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.275870] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.276193] env[63418]: DEBUG nova.compute.manager [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 922.276524] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 922.276524] env[63418]: value = "task-1245247" [ 922.276524] env[63418]: _type = "Task" [ 922.276524] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.277229] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edbbce7-f8b1-4e62-8891-636a073364cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.290866] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.303251] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245244, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.745324} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.303478] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 4838d2b2-a187-48ae-a3b1-8ca24b961359/4838d2b2-a187-48ae-a3b1-8ca24b961359.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.303703] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.303940] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07b7cb7c-9a09-47cf-b848-24b6d8908c74 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.311670] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 922.311670] env[63418]: value = "task-1245248" [ 922.311670] env[63418]: _type = "Task" [ 922.311670] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.319631] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245248, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.388272] env[63418]: DEBUG nova.network.neutron [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Updated VIF entry in instance network info cache for port fc06d529-525c-4dbf-8050-32c6054ecc96. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.388655] env[63418]: DEBUG nova.network.neutron [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Updating instance_info_cache with network_info: [{"id": "fc06d529-525c-4dbf-8050-32c6054ecc96", "address": "fa:16:3e:b5:18:2e", "network": {"id": "bb6fe8e2-40b2-4a1b-917f-8c30c020ce2c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1142963876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4759c1d6199344c49063d35e422d6324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc06d529-52", "ovs_interfaceid": "fc06d529-525c-4dbf-8050-32c6054ecc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.493376] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.987s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.493940] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 922.496841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.755s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.498694] env[63418]: INFO nova.compute.claims [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.543681] env[63418]: DEBUG oslo_vmware.api [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245245, 'name': ReconfigVM_Task, 'duration_secs': 0.204611} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.543985] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Reconfigured VM instance instance-00000031 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 922.549397] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c4fbe10-c505-4433-85cd-d30ca1d022be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.564981] env[63418]: DEBUG oslo_vmware.api [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 922.564981] env[63418]: value = "task-1245249" [ 922.564981] env[63418]: _type = "Task" [ 922.564981] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.574315] env[63418]: DEBUG oslo_vmware.api [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.757797] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245246, 'name': Rename_Task, 'duration_secs': 0.158467} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.758094] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.758346] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5c05edd-2efb-45fd-ad27-d0bd9ce4dc83 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.764403] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 922.764403] env[63418]: value = "task-1245250" [ 922.764403] env[63418]: _type = "Task" [ 922.764403] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.772736] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.788740] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073219} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.789154] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.789923] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb6d7af-c695-4772-bc09-754a2457b4d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.796358] env[63418]: INFO nova.compute.manager [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Shelve offloading [ 922.818950] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/24a97d0a-252b-4bbd-9dfc-57767ef8014d.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.820775] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-846f31b6-e8ff-4abf-ad94-9ef0ac799198 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.846067] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245248, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068103} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.847784] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.848174] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 922.848174] env[63418]: value = "task-1245251" [ 922.848174] env[63418]: _type = "Task" [ 922.848174] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.848860] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eaa9cad-ef89-407c-b01c-400860884c9b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.860124] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245251, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.880967] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 4838d2b2-a187-48ae-a3b1-8ca24b961359/4838d2b2-a187-48ae-a3b1-8ca24b961359.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.881370] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26f788af-b54f-4936-8135-a259f89d16a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.899227] env[63418]: DEBUG oslo_concurrency.lockutils [req-e302cc0b-ed86-4436-994f-39a80b3c41cb req-ff1ea924-511f-47ec-97c6-3c46cf4fcf79 service nova] Releasing lock "refresh_cache-4838d2b2-a187-48ae-a3b1-8ca24b961359" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.905451] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 922.905451] env[63418]: value = "task-1245252" [ 922.905451] env[63418]: _type = "Task" [ 922.905451] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.913928] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245252, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.008050] env[63418]: DEBUG nova.compute.utils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.009473] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 923.009603] env[63418]: DEBUG nova.network.neutron [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 923.071111] env[63418]: DEBUG nova.policy [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94b1aab9fe174e8caa3a809e1b642131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aff8e2c3d27c4277b2b8f130c293c013', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.078471] env[63418]: DEBUG oslo_vmware.api [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245249, 'name': ReconfigVM_Task, 'duration_secs': 0.188462} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.078653] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268463', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'name': 'volume-ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'd76a008c-9bd9-420b-873d-4f7d7f25b8ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967', 'serial': 'ea6c4a8b-acab-4ea9-a5d5-5b1f0cdb1967'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 923.275671] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245250, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.338063] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.338418] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dad8c597-b609-4f03-afbb-85847d1f5c0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.345040] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 923.345040] env[63418]: value = "task-1245253" [ 923.345040] env[63418]: _type = "Task" [ 923.345040] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.352996] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.360087] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245251, 'name': ReconfigVM_Task, 'duration_secs': 0.307644} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.360424] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/24a97d0a-252b-4bbd-9dfc-57767ef8014d.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.361075] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad577b4c-ce2e-4720-a49e-ff39dd95fc7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.367529] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 923.367529] env[63418]: value = "task-1245254" [ 923.367529] env[63418]: _type = "Task" [ 923.367529] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.377166] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245254, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.401328] env[63418]: DEBUG nova.network.neutron [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Successfully created port: 01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.415435] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245252, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.473032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.473032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.473032] env[63418]: DEBUG nova.compute.manager [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Going to confirm migration 1 {{(pid=63418) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 923.513268] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 923.628518] env[63418]: DEBUG nova.objects.instance [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'flavor' on Instance uuid d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.777249] env[63418]: DEBUG oslo_vmware.api [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245250, 'name': PowerOnVM_Task, 'duration_secs': 0.512494} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.777589] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.778204] env[63418]: INFO nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Took 11.23 seconds to spawn the instance on the hypervisor. [ 923.778204] env[63418]: DEBUG nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 923.778906] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3953229-2552-451c-ab2f-ee6d1996bd86 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.784835] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac81b75-adaf-4732-88af-26a5bc189b64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.796024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49344ddd-7eab-4e3e-a637-a336fb7704fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.831320] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884a1098-abff-4993-8a3c-602b84a0cc82 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.839628] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87570de1-2ae9-4a1e-9923-91c3800ca818 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.856206] env[63418]: DEBUG nova.compute.provider_tree [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.862826] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 923.863069] env[63418]: DEBUG nova.compute.manager [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 923.863827] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c77b080-7abf-4f6a-a698-f0f1d820b7ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.869938] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.870646] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.870865] env[63418]: DEBUG nova.network.neutron [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.881302] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245254, 'name': Rename_Task, 'duration_secs': 0.14122} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.881557] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.881784] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-337c9315-b0ce-40e2-a8c8-359118f91583 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.888550] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 923.888550] env[63418]: value = "task-1245255" [ 923.888550] env[63418]: _type = "Task" [ 923.888550] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.895944] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245255, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.914007] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245252, 'name': ReconfigVM_Task, 'duration_secs': 0.674764} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.914701] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 4838d2b2-a187-48ae-a3b1-8ca24b961359/4838d2b2-a187-48ae-a3b1-8ca24b961359.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.915367] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c578f73-2726-481a-9406-3195f690040e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.923394] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 923.923394] env[63418]: value = "task-1245256" [ 923.923394] env[63418]: _type = "Task" [ 923.923394] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.931507] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245256, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.066240] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.066462] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.066716] env[63418]: DEBUG nova.network.neutron [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.066927] env[63418]: DEBUG nova.objects.instance [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lazy-loading 'info_cache' on Instance uuid 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.300560] env[63418]: INFO nova.compute.manager [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Took 19.99 seconds to build instance. [ 924.361061] env[63418]: DEBUG nova.scheduler.client.report [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 924.399133] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245255, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.432414] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245256, 'name': Rename_Task, 'duration_secs': 0.242215} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.432707] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.433037] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7bb66cf2-0dbb-4fb0-bd0f-e7eeba1c7d77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.440691] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 924.440691] env[63418]: value = "task-1245257" [ 924.440691] env[63418]: _type = "Task" [ 924.440691] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.450803] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245257, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.527036] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 924.555874] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.556165] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.556364] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.556603] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.556765] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.556920] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.557153] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.557324] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.557502] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.557671] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.557847] env[63418]: DEBUG nova.virt.hardware [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.558724] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726daf8d-d199-4917-a6c7-1b2171adb1cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.567168] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc31f9dd-5f19-4b17-9cbc-83cf4992e03f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.587550] env[63418]: DEBUG nova.network.neutron [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.637191] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0bd5c4ab-78ba-45a8-a758-e04bc3a9a31f tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.340s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.803208] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1a771635-8541-4854-8e55-5e86e55b8bdf tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.496s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.864373] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.864929] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 924.867835] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.452s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.906337] env[63418]: DEBUG oslo_vmware.api [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245255, 'name': PowerOnVM_Task, 'duration_secs': 0.648503} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.906864] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.907155] env[63418]: INFO nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Took 9.91 seconds to spawn the instance on the hypervisor. [ 924.907445] env[63418]: DEBUG nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 924.908341] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b170ec-54bf-4ba1-9a55-1f132d68de49 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.934366] env[63418]: DEBUG nova.network.neutron [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Successfully updated port: 01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.953224] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245257, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.974524] env[63418]: DEBUG nova.compute.manager [req-7dec35a1-e551-40f9-a85c-7e3a786d8d6c req-21fb8358-2846-4520-a27a-01e417431a82 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Received event network-vif-plugged-01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 924.974776] env[63418]: DEBUG oslo_concurrency.lockutils [req-7dec35a1-e551-40f9-a85c-7e3a786d8d6c req-21fb8358-2846-4520-a27a-01e417431a82 service nova] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.974995] env[63418]: DEBUG oslo_concurrency.lockutils [req-7dec35a1-e551-40f9-a85c-7e3a786d8d6c req-21fb8358-2846-4520-a27a-01e417431a82 service nova] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.975195] env[63418]: DEBUG oslo_concurrency.lockutils [req-7dec35a1-e551-40f9-a85c-7e3a786d8d6c req-21fb8358-2846-4520-a27a-01e417431a82 service nova] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.975366] env[63418]: DEBUG nova.compute.manager [req-7dec35a1-e551-40f9-a85c-7e3a786d8d6c req-21fb8358-2846-4520-a27a-01e417431a82 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] No waiting events found dispatching network-vif-plugged-01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 924.975536] env[63418]: WARNING nova.compute.manager [req-7dec35a1-e551-40f9-a85c-7e3a786d8d6c req-21fb8358-2846-4520-a27a-01e417431a82 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Received unexpected event network-vif-plugged-01df43f5-4069-46a0-b1e6-4baff131e04e for instance with vm_state building and task_state spawning. [ 925.090700] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.373399] env[63418]: DEBUG nova.compute.utils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.378009] env[63418]: INFO nova.compute.claims [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.385864] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 925.385864] env[63418]: DEBUG nova.network.neutron [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 925.425632] env[63418]: DEBUG nova.policy [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.433507] env[63418]: INFO nova.compute.manager [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Took 16.69 seconds to build instance. [ 925.437454] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.437632] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.437952] env[63418]: DEBUG nova.network.neutron [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.444116] env[63418]: DEBUG nova.network.neutron [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance_info_cache with network_info: [{"id": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "address": "fa:16:3e:72:aa:e2", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ff1dd4-2e", "ovs_interfaceid": "e4ff1dd4-2e2f-42cb-b882-19c984a2eb28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.454458] env[63418]: DEBUG oslo_vmware.api [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245257, 'name': PowerOnVM_Task, 'duration_secs': 0.704582} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.454895] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.455299] env[63418]: INFO nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Took 8.09 seconds to spawn the instance on the hypervisor. [ 925.457998] env[63418]: DEBUG nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 925.457998] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a52ada-9c26-4b36-8cd3-5498385daff2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.571236] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.571236] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f7fcbb-d029-4d42-a19a-034891080ff0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.578990] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.579440] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8accc65e-f2ff-4773-b0cd-a638f07c64a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.658111] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.658111] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.658111] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleting the datastore file [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.658111] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d93870f-ac46-4007-958d-34bb484a94dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.663876] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 925.663876] env[63418]: value = "task-1245259" [ 925.663876] env[63418]: _type = "Task" [ 925.663876] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.671969] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245259, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.884145] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 925.889643] env[63418]: INFO nova.compute.resource_tracker [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating resource usage from migration 35dd30ef-05a7-46f8-bed4-fd19eb2c173a [ 925.935974] env[63418]: DEBUG oslo_concurrency.lockutils [None req-24a528ed-5aaa-4e23-bc3d-ef3567b131fc tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.206s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.950872] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.950872] env[63418]: DEBUG nova.objects.instance [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lazy-loading 'migration_context' on Instance uuid 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.962348] env[63418]: DEBUG nova.network.neutron [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Successfully created port: 34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.976968] env[63418]: INFO nova.compute.manager [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Took 14.32 seconds to build instance. [ 925.996593] env[63418]: DEBUG nova.network.neutron [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.158134] env[63418]: DEBUG nova.network.neutron [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [{"id": "01df43f5-4069-46a0-b1e6-4baff131e04e", "address": "fa:16:3e:69:39:54", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01df43f5-40", "ovs_interfaceid": "01df43f5-4069-46a0-b1e6-4baff131e04e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.178132] env[63418]: DEBUG oslo_vmware.api [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245259, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181421} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.179116] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.179464] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.179651] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.193920] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25ce0bd-5270-4b5c-9330-59a6d6c57af2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.198995] env[63418]: INFO nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted allocations for instance 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 [ 926.206048] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f981bef-7874-4b9a-bf97-f2cdb3b19f44 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.234438] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d498116-edea-4ceb-83ff-54626f84331d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.243118] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73d086e-f2fc-4bbf-875a-dba1ae62cda1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.258946] env[63418]: DEBUG nova.compute.provider_tree [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.431844] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.432223] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.432486] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.432681] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.432858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.435418] env[63418]: INFO nova.compute.manager [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Terminating instance [ 926.449026] env[63418]: INFO nova.compute.manager [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Rescuing [ 926.449026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.449026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.449026] env[63418]: DEBUG nova.network.neutron [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 926.453211] env[63418]: DEBUG nova.objects.base [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Object Instance<1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5> lazy-loaded attributes: info_cache,migration_context {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 926.454388] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ee7618-3e25-4622-9cf2-f8e0f9ada6ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.476281] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42db58cb-648a-4cde-a21c-fb6a49137a96 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.480496] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7c78f185-8d69-444b-9c6d-d2b5f93f5208 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.837s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.482405] env[63418]: DEBUG oslo_vmware.api [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 926.482405] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5222c1c6-46e7-0e5e-bec5-d462e7424a5e" [ 926.482405] env[63418]: _type = "Task" [ 926.482405] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.492211] env[63418]: DEBUG oslo_vmware.api [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5222c1c6-46e7-0e5e-bec5-d462e7424a5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.507757] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "14832e8e-3b4a-462e-84bb-d5a1e541d329" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.508302] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.508614] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "14832e8e-3b4a-462e-84bb-d5a1e541d329-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.508858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.509054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.512837] env[63418]: INFO nova.compute.manager [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Terminating instance [ 926.661681] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.662058] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Instance network_info: |[{"id": "01df43f5-4069-46a0-b1e6-4baff131e04e", "address": "fa:16:3e:69:39:54", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01df43f5-40", "ovs_interfaceid": "01df43f5-4069-46a0-b1e6-4baff131e04e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 926.662730] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:39:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01df43f5-4069-46a0-b1e6-4baff131e04e', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.675048] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Creating folder: Project (aff8e2c3d27c4277b2b8f130c293c013). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 926.675420] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ef6f659-12d8-421c-b225-6316bc654112 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.686554] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Created folder: Project (aff8e2c3d27c4277b2b8f130c293c013) in parent group-v268354. [ 926.686792] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Creating folder: Instances. Parent ref: group-v268504. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 926.687087] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-066370d2-64d2-4e14-a196-54a50404174c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.695810] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Created folder: Instances in parent group-v268504. [ 926.696117] env[63418]: DEBUG oslo.service.loopingcall [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.696359] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.696601] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f66dd65e-68f8-4465-bbf4-3b04501aefa7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.719788] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.725231] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.725231] env[63418]: value = "task-1245262" [ 926.725231] env[63418]: _type = "Task" [ 926.725231] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.734698] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245262, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.763837] env[63418]: DEBUG nova.scheduler.client.report [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 926.904852] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 926.931910] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.932201] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.932385] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.932578] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.932733] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.932888] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.933483] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.933707] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.933938] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.934175] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.934397] env[63418]: DEBUG nova.virt.hardware [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.935622] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471af55b-2bf1-4ece-becd-95d09a8ae19f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.940017] env[63418]: DEBUG nova.compute.manager [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 926.940017] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.942553] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992830f0-9fee-4239-827f-d7d5e9b3c6a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.956062] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73b77c6-42a4-47a8-aceb-4a3a89873d90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.960097] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.960601] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4d48547-e433-47b7-af6d-f8f9f071f25a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.974041] env[63418]: DEBUG oslo_vmware.api [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 926.974041] env[63418]: value = "task-1245263" [ 926.974041] env[63418]: _type = "Task" [ 926.974041] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.982310] env[63418]: DEBUG oslo_vmware.api [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.991978] env[63418]: DEBUG oslo_vmware.api [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5222c1c6-46e7-0e5e-bec5-d462e7424a5e, 'name': SearchDatastore_Task, 'duration_secs': 0.009865} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.992337] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.008407] env[63418]: DEBUG nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Received event network-changed-01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 927.008600] env[63418]: DEBUG nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Refreshing instance network info cache due to event network-changed-01df43f5-4069-46a0-b1e6-4baff131e04e. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 927.008800] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Acquiring lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.009574] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Acquired lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.009574] env[63418]: DEBUG nova.network.neutron [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Refreshing network info cache for port 01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.017203] env[63418]: DEBUG nova.compute.manager [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 927.017425] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.019209] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e7f83f-bc93-45d0-84d4-eb7df4434fab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.027708] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.027997] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6658ab0e-f7fc-46e9-baf3-60a32899b82f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.038864] env[63418]: DEBUG oslo_vmware.api [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 927.038864] env[63418]: value = "task-1245264" [ 927.038864] env[63418]: _type = "Task" [ 927.038864] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.047987] env[63418]: DEBUG oslo_vmware.api [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.179885] env[63418]: DEBUG nova.network.neutron [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updating instance_info_cache with network_info: [{"id": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "address": "fa:16:3e:36:3a:6f", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef27d89f-f8", "ovs_interfaceid": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.234305] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245262, 'name': CreateVM_Task, 'duration_secs': 0.475355} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.234505] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.235365] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.235524] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.235807] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.236089] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c952f0f8-21ce-45fd-9987-463ca0482e25 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.241150] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 927.241150] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5246bc6a-154a-f692-4262-fb941db46908" [ 927.241150] env[63418]: _type = "Task" [ 927.241150] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.249891] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5246bc6a-154a-f692-4262-fb941db46908, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.268860] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.401s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.269098] env[63418]: INFO nova.compute.manager [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Migrating [ 927.275683] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.514s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.277591] env[63418]: INFO nova.compute.claims [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.307231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "4838d2b2-a187-48ae-a3b1-8ca24b961359" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.307231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.307231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "4838d2b2-a187-48ae-a3b1-8ca24b961359-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.307231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.307231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.308047] env[63418]: INFO nova.compute.manager [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Terminating instance [ 927.488290] env[63418]: DEBUG oslo_vmware.api [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245263, 'name': PowerOffVM_Task, 'duration_secs': 0.352709} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.488582] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.488752] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.489009] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5335af25-de2f-4a5d-a93e-cda75b3268fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.552942] env[63418]: DEBUG oslo_vmware.api [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245264, 'name': PowerOffVM_Task, 'duration_secs': 0.290413} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.553768] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.553950] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.554228] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.554435] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.554619] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleting the datastore file [datastore1] d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.554861] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5109cd95-3533-4c8e-bc82-08f4c4e7f0de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.556367] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41b21a1b-f0cb-4ce3-acb3-958efdaa2e3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.564165] env[63418]: DEBUG oslo_vmware.api [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 927.564165] env[63418]: value = "task-1245266" [ 927.564165] env[63418]: _type = "Task" [ 927.564165] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.569956] env[63418]: DEBUG oslo_vmware.api [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.628089] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.628351] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.628608] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore1] 14832e8e-3b4a-462e-84bb-d5a1e541d329 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.628756] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad0d2836-1b74-44f2-9f57-98c8d560c049 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.635149] env[63418]: DEBUG oslo_vmware.api [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 927.635149] env[63418]: value = "task-1245268" [ 927.635149] env[63418]: _type = "Task" [ 927.635149] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.646219] env[63418]: DEBUG oslo_vmware.api [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245268, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.683928] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.741067] env[63418]: DEBUG nova.network.neutron [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updated VIF entry in instance network info cache for port 01df43f5-4069-46a0-b1e6-4baff131e04e. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.741456] env[63418]: DEBUG nova.network.neutron [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [{"id": "01df43f5-4069-46a0-b1e6-4baff131e04e", "address": "fa:16:3e:69:39:54", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01df43f5-40", "ovs_interfaceid": "01df43f5-4069-46a0-b1e6-4baff131e04e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.753771] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5246bc6a-154a-f692-4262-fb941db46908, 'name': SearchDatastore_Task, 'duration_secs': 0.008986} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.754782] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.755058] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.755335] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.755488] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.755710] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.756274] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c97a4c39-fe89-4d01-8eb0-e711b188ee39 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.773755] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.773975] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.774815] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cedb89af-e42a-4ff4-9d63-ee547afec302 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.783759] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 927.783759] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5225b7e0-7995-bcb6-8414-5d430219cc35" [ 927.783759] env[63418]: _type = "Task" [ 927.783759] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.789858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.789858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.789858] env[63418]: DEBUG nova.network.neutron [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.793231] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5225b7e0-7995-bcb6-8414-5d430219cc35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.795174] env[63418]: DEBUG nova.compute.manager [req-3a0dddbe-ff6f-4e58-a5b3-5fee150cc99f req-cad400cb-908b-48b2-8280-086365cb7d30 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-vif-plugged-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 927.795377] env[63418]: DEBUG oslo_concurrency.lockutils [req-3a0dddbe-ff6f-4e58-a5b3-5fee150cc99f req-cad400cb-908b-48b2-8280-086365cb7d30 service nova] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.795569] env[63418]: DEBUG oslo_concurrency.lockutils [req-3a0dddbe-ff6f-4e58-a5b3-5fee150cc99f req-cad400cb-908b-48b2-8280-086365cb7d30 service nova] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.795739] env[63418]: DEBUG oslo_concurrency.lockutils [req-3a0dddbe-ff6f-4e58-a5b3-5fee150cc99f req-cad400cb-908b-48b2-8280-086365cb7d30 service nova] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.795907] env[63418]: DEBUG nova.compute.manager [req-3a0dddbe-ff6f-4e58-a5b3-5fee150cc99f req-cad400cb-908b-48b2-8280-086365cb7d30 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] No waiting events found dispatching network-vif-plugged-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 927.796091] env[63418]: WARNING nova.compute.manager [req-3a0dddbe-ff6f-4e58-a5b3-5fee150cc99f req-cad400cb-908b-48b2-8280-086365cb7d30 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received unexpected event network-vif-plugged-34d14910-009f-4ee8-b718-a43961c430cb for instance with vm_state building and task_state spawning. [ 927.811831] env[63418]: DEBUG nova.compute.manager [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 927.812080] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.813016] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a72aca-d12a-435c-9a9f-661c9f34bbcf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.821811] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.822588] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7a2cd0a-963e-45d2-8cc0-53dba2963d19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.828518] env[63418]: DEBUG oslo_vmware.api [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 927.828518] env[63418]: value = "task-1245269" [ 927.828518] env[63418]: _type = "Task" [ 927.828518] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.837208] env[63418]: DEBUG oslo_vmware.api [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245269, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.002679] env[63418]: DEBUG nova.network.neutron [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Successfully updated port: 34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.075528] env[63418]: DEBUG oslo_vmware.api [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219849} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.075961] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.076423] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.076780] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.077056] env[63418]: INFO nova.compute.manager [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Took 1.14 seconds to destroy the instance on the hypervisor. [ 928.077658] env[63418]: DEBUG oslo.service.loopingcall [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.077786] env[63418]: DEBUG nova.compute.manager [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 928.078058] env[63418]: DEBUG nova.network.neutron [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.145992] env[63418]: DEBUG oslo_vmware.api [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245268, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18782} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.148671] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.148897] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.149128] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.149449] env[63418]: INFO nova.compute.manager [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Took 1.13 seconds to destroy the instance on the hypervisor. [ 928.150054] env[63418]: DEBUG oslo.service.loopingcall [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.150298] env[63418]: DEBUG nova.compute.manager [-] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 928.150396] env[63418]: DEBUG nova.network.neutron [-] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.244605] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Releasing lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.244881] env[63418]: DEBUG nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-vif-unplugged-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 928.245085] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.245318] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.245531] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.245707] env[63418]: DEBUG nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] No waiting events found dispatching network-vif-unplugged-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 928.245881] env[63418]: WARNING nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received unexpected event network-vif-unplugged-f786c95b-5214-454d-86f8-6d922f0482d8 for instance with vm_state shelved_offloaded and task_state None. [ 928.246062] env[63418]: DEBUG nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-changed-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 928.246228] env[63418]: DEBUG nova.compute.manager [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Refreshing instance network info cache due to event network-changed-f786c95b-5214-454d-86f8-6d922f0482d8. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 928.246422] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.246565] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.246764] env[63418]: DEBUG nova.network.neutron [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Refreshing network info cache for port f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.301108] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5225b7e0-7995-bcb6-8414-5d430219cc35, 'name': SearchDatastore_Task, 'duration_secs': 0.010573} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.302696] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b6eb6da-7d3f-4271-b0f9-453058937377 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.309550] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 928.309550] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e5e681-29ac-0730-b3e1-8c00c3cb891e" [ 928.309550] env[63418]: _type = "Task" [ 928.309550] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.322106] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e5e681-29ac-0730-b3e1-8c00c3cb891e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.337628] env[63418]: DEBUG oslo_vmware.api [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245269, 'name': PowerOffVM_Task, 'duration_secs': 0.483506} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.340321] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.340527] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.340970] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c389296-b8e5-4214-a199-d57076d46f06 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.405669] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.405669] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.405849] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleting the datastore file [datastore1] 4838d2b2-a187-48ae-a3b1-8ca24b961359 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.406088] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bf2ceea-17d2-482c-aa8d-8117c7413e4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.414633] env[63418]: DEBUG oslo_vmware.api [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for the task: (returnval){ [ 928.414633] env[63418]: value = "task-1245271" [ 928.414633] env[63418]: _type = "Task" [ 928.414633] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.422531] env[63418]: DEBUG oslo_vmware.api [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.506089] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.506089] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.506089] env[63418]: DEBUG nova.network.neutron [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 928.579869] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e7eaba-a776-4a2a-b0f8-6bdee9a739bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.587359] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced6c749-92ca-4952-b453-b307454a21b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.617139] env[63418]: DEBUG nova.network.neutron [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.618878] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35eac9bf-dd2a-4d2e-bd5c-9459b7bed914 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.626911] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d6f955-5929-4ee8-ad1e-c01b85efc578 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.643062] env[63418]: DEBUG nova.compute.provider_tree [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.726457] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.726777] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf6d7d86-c5f8-4476-b188-d626dc979e1e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.736595] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 928.736595] env[63418]: value = "task-1245272" [ 928.736595] env[63418]: _type = "Task" [ 928.736595] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.746042] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.826029] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e5e681-29ac-0730-b3e1-8c00c3cb891e, 'name': SearchDatastore_Task, 'duration_secs': 0.01831} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.826029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.826029] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] cbcbaf4f-ee24-4072-83ae-ffde59478928/cbcbaf4f-ee24-4072-83ae-ffde59478928.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.826029] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76482086-fb48-4ba7-8e38-7b18878abdb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.838021] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 928.838021] env[63418]: value = "task-1245273" [ 928.838021] env[63418]: _type = "Task" [ 928.838021] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.845916] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245273, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.924198] env[63418]: DEBUG oslo_vmware.api [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Task: {'id': task-1245271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369005} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.924739] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.925107] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.925419] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.925765] env[63418]: INFO nova.compute.manager [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Took 1.11 seconds to destroy the instance on the hypervisor. [ 928.926147] env[63418]: DEBUG oslo.service.loopingcall [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.926874] env[63418]: DEBUG nova.compute.manager [-] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 928.927102] env[63418]: DEBUG nova.network.neutron [-] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.978160] env[63418]: DEBUG nova.network.neutron [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updated VIF entry in instance network info cache for port f786c95b-5214-454d-86f8-6d922f0482d8. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.978160] env[63418]: DEBUG nova.network.neutron [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": null, "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf786c95b-52", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.993164] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.999023] env[63418]: DEBUG nova.network.neutron [-] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.037300] env[63418]: DEBUG nova.compute.manager [req-e0f9750c-2722-43c6-b24e-c6fe8353f7ee req-787b8c2d-5f5c-40c7-bd64-5c54c692e5f8 service nova] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Received event network-vif-deleted-4fdaa32d-3674-4689-8a75-07aa2e067de4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 929.046324] env[63418]: DEBUG nova.network.neutron [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 929.122771] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.148616] env[63418]: DEBUG nova.scheduler.client.report [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 929.241588] env[63418]: DEBUG nova.network.neutron [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.254339] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245272, 'name': PowerOffVM_Task, 'duration_secs': 0.28786} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.254681] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.255605] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b807d41c-97fb-4920-8a55-4edd437c3cc5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.285757] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b1a8d8-74bc-4131-a4b3-ee59b4a92448 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.324887] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.325599] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fcbe311-cf3f-4199-9784-aa3b85f725f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.334014] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 929.334014] env[63418]: value = "task-1245274" [ 929.334014] env[63418]: _type = "Task" [ 929.334014] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.345935] env[63418]: DEBUG nova.network.neutron [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.353863] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 929.354118] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.354414] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.354572] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.354807] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.358660] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d927ded-f77b-453d-8822-76f93e850ae3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.360812] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245273, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.368100] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.368393] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.369976] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a65c7bda-b209-4ef2-87f1-bda253dad723 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.378675] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 929.378675] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52583a91-1f96-76ab-ebb3-bc2dee24036c" [ 929.378675] env[63418]: _type = "Task" [ 929.378675] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.385713] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52583a91-1f96-76ab-ebb3-bc2dee24036c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.479523] env[63418]: DEBUG oslo_concurrency.lockutils [req-789c88a4-81ff-43a9-a082-435b50404fc5 req-57fb3aad-ce53-4588-9a67-5567119b7673 service nova] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.501549] env[63418]: INFO nova.compute.manager [-] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Took 1.35 seconds to deallocate network for instance. [ 929.652039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.652501] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 929.655050] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.875s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.656521] env[63418]: INFO nova.compute.claims [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.694566] env[63418]: DEBUG nova.network.neutron [-] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.745270] env[63418]: INFO nova.compute.manager [-] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Took 1.67 seconds to deallocate network for instance. [ 929.848337] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245273, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655376} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.848583] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] cbcbaf4f-ee24-4072-83ae-ffde59478928/cbcbaf4f-ee24-4072-83ae-ffde59478928.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.848786] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.849045] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d0226ac-2e50-4942-92de-e3bf75d94044 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.854670] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.855290] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Instance network_info: |[{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 929.856453] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:7f:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34d14910-009f-4ee8-b718-a43961c430cb', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.865290] env[63418]: DEBUG oslo.service.loopingcall [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.865290] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 929.865290] env[63418]: value = "task-1245275" [ 929.865290] env[63418]: _type = "Task" [ 929.865290] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.865574] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.865697] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c76c0bc3-0b32-4bd6-859c-3037787774ef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.885506] env[63418]: DEBUG nova.compute.manager [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-changed-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 929.885655] env[63418]: DEBUG nova.compute.manager [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing instance network info cache due to event network-changed-34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 929.886032] env[63418]: DEBUG oslo_concurrency.lockutils [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.886032] env[63418]: DEBUG oslo_concurrency.lockutils [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.886244] env[63418]: DEBUG nova.network.neutron [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.897598] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245275, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.902558] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52583a91-1f96-76ab-ebb3-bc2dee24036c, 'name': SearchDatastore_Task, 'duration_secs': 0.021325} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.902993] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.902993] env[63418]: value = "task-1245276" [ 929.902993] env[63418]: _type = "Task" [ 929.902993] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.903962] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f499b3a-0784-40d9-9ea5-c60ee1933790 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.918274] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245276, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.919664] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 929.919664] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529190f4-7389-1347-0a67-e9cb3f6bfcd3" [ 929.919664] env[63418]: _type = "Task" [ 929.919664] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.929102] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529190f4-7389-1347-0a67-e9cb3f6bfcd3, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.929373] env[63418]: DEBUG oslo_concurrency.lockutils [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.929634] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk. {{(pid=63418) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 929.929888] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8b02652-d75b-4d7d-9d38-6c328e27e56e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.936246] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 929.936246] env[63418]: value = "task-1245277" [ 929.936246] env[63418]: _type = "Task" [ 929.936246] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.944463] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.009424] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.160347] env[63418]: DEBUG nova.compute.utils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.164334] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 930.164573] env[63418]: DEBUG nova.network.neutron [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.197108] env[63418]: INFO nova.compute.manager [-] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Took 1.27 seconds to deallocate network for instance. [ 930.224870] env[63418]: DEBUG nova.policy [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ff64612530b451fb41100a5aa601be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d52f1fceb24234a8d967038b43c857', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.251960] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.375928] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245275, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060239} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.376305] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.377118] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cfe600-cca2-41fe-ace1-e9411ea171d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.414569] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] cbcbaf4f-ee24-4072-83ae-ffde59478928/cbcbaf4f-ee24-4072-83ae-ffde59478928.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.415888] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7767d6f-9db1-4d71-9704-94b5f6501284 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.445804] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 930.445804] env[63418]: value = "task-1245278" [ 930.445804] env[63418]: _type = "Task" [ 930.445804] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.451844] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436051} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.452114] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245276, 'name': CreateVM_Task, 'duration_secs': 0.466994} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.455201] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk. [ 930.455462] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.456197] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8a6718-4a36-4845-b077-170ddb116d50 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.459136] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.459377] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.459737] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.460397] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e65ec25-21f2-4c64-b771-3486248d07f5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.465247] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.486755] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.490276] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7900fa4c-13af-49c8-9162-42c4c1ab20a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.503294] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 930.503294] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aac1c7-f9f1-9e3d-667b-02cda1f373ed" [ 930.503294] env[63418]: _type = "Task" [ 930.503294] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.511621] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 930.511621] env[63418]: value = "task-1245279" [ 930.511621] env[63418]: _type = "Task" [ 930.511621] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.519192] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52aac1c7-f9f1-9e3d-667b-02cda1f373ed, 'name': SearchDatastore_Task, 'duration_secs': 0.009475} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.519852] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.520124] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.520359] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.520550] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.520777] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.520986] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fdba7bd-1bd6-438a-8fc0-d906b494d15f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.526131] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245279, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.532313] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.532503] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.533302] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d30020a-b0aa-4313-8618-4e649fe53b32 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.538197] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 930.538197] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c12bff-d921-f429-0574-bf703e5a0be3" [ 930.538197] env[63418]: _type = "Task" [ 930.538197] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.545747] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c12bff-d921-f429-0574-bf703e5a0be3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.642777] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df04e89c-026c-48c3-8bf3-a9d094e5e391 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.662094] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 0 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 930.666114] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 930.704529] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.717768] env[63418]: DEBUG nova.network.neutron [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Successfully created port: 1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.732796] env[63418]: DEBUG nova.network.neutron [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updated VIF entry in instance network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.733207] env[63418]: DEBUG nova.network.neutron [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.938225] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d480f5-bcfc-43d4-9de4-864633983ed0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.946675] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f58944-8aa0-4878-bb50-048c9cee5243 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.979721] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf8f5fb-2595-489b-ac72-9b843ecaf868 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.985237] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245278, 'name': ReconfigVM_Task, 'duration_secs': 0.306901} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.985868] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Reconfigured VM instance instance-00000058 to attach disk [datastore2] cbcbaf4f-ee24-4072-83ae-ffde59478928/cbcbaf4f-ee24-4072-83ae-ffde59478928.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.986534] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7bc9b09-921a-43ad-ac16-7e3b20b85224 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.991194] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9433b9-b9e8-4945-8f71-2cbe404d1223 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.997747] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 930.997747] env[63418]: value = "task-1245280" [ 930.997747] env[63418]: _type = "Task" [ 930.997747] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.008546] env[63418]: DEBUG nova.compute.provider_tree [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.014520] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245280, 'name': Rename_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.023640] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245279, 'name': ReconfigVM_Task, 'duration_secs': 0.330266} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.023925] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d/c0dccf29-5e49-4a1e-b51f-d46e566b4772-rescue.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.024830] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90d289e-20f3-40ba-9a60-4def898680bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.053230] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3d593d2-872e-48dd-a073-3f00b8602ec7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.070589] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c12bff-d921-f429-0574-bf703e5a0be3, 'name': SearchDatastore_Task, 'duration_secs': 0.010156} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.072884] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 931.072884] env[63418]: value = "task-1245281" [ 931.072884] env[63418]: _type = "Task" [ 931.072884] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.073187] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fbed373-9ce2-4992-84e5-53b42ef5a8f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.084234] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245281, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.085308] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 931.085308] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5242a640-1926-7838-51ac-4c3550265f60" [ 931.085308] env[63418]: _type = "Task" [ 931.085308] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.092878] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5242a640-1926-7838-51ac-4c3550265f60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.172882] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.176317] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68a5e4d4-4ea1-4889-8b03-d2adbcb1eacd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.186015] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 931.186015] env[63418]: value = "task-1245282" [ 931.186015] env[63418]: _type = "Task" [ 931.186015] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.191334] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245282, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.238325] env[63418]: DEBUG oslo_concurrency.lockutils [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.238664] env[63418]: DEBUG nova.compute.manager [req-455fb1c6-d0e2-4d42-873d-f5027c51eaa6 req-36c2aa43-4fdb-4e37-98cf-d1f7b7111d4e service nova] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Received event network-vif-deleted-0263ccbe-8541-4cf4-bd2a-0e9b517d6f29 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 931.247307] env[63418]: DEBUG nova.compute.manager [req-7231d9cc-e278-483b-b82e-f4877780252b req-7b28895b-25d8-44a2-8a7d-243972b615ab service nova] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Received event network-vif-deleted-fc06d529-525c-4dbf-8050-32c6054ecc96 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 931.516304] env[63418]: DEBUG nova.scheduler.client.report [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 931.525590] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245280, 'name': Rename_Task, 'duration_secs': 0.152912} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.526490] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.526490] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e29a40f-1b4c-4f99-8bd5-3f187242bf2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.533547] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 931.533547] env[63418]: value = "task-1245283" [ 931.533547] env[63418]: _type = "Task" [ 931.533547] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.541310] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.585667] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245281, 'name': ReconfigVM_Task, 'duration_secs': 0.218217} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.585667] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.585768] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e41874e0-0ca1-46fa-93d2-feef04f793d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.602220] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5242a640-1926-7838-51ac-4c3550265f60, 'name': SearchDatastore_Task, 'duration_secs': 0.012085} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.602220] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.602220] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0f99b32a-0125-4df0-919c-e5456b4ae4a5/0f99b32a-0125-4df0-919c-e5456b4ae4a5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.602220] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-207c3e79-982c-4388-9276-6feef2ce489b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.607585] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 931.607585] env[63418]: value = "task-1245284" [ 931.607585] env[63418]: _type = "Task" [ 931.607585] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.612340] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 931.612340] env[63418]: value = "task-1245285" [ 931.612340] env[63418]: _type = "Task" [ 931.612340] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.629153] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245285, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.629528] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245284, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.680958] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 931.694728] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 931.695049] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 17 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 931.733248] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.733521] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.733700] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.733842] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.733994] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.734385] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.735112] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.735112] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.735112] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.735112] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.735380] env[63418]: DEBUG nova.virt.hardware [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.736218] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fbbe6e-ec9d-4c08-b8ce-64798748c69c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.745829] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440706df-adc1-406a-9319-ff179ddf60a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.027996] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.028867] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 932.032270] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.313s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.032520] env[63418]: DEBUG nova.objects.instance [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lazy-loading 'resources' on Instance uuid 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.044242] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245283, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.118090] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245284, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.126277] env[63418]: DEBUG oslo_vmware.api [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245285, 'name': PowerOnVM_Task, 'duration_secs': 0.489209} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.126546] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.129217] env[63418]: DEBUG nova.compute.manager [None req-25179308-d326-4d3c-b60c-cccf7bd56751 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 932.129968] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd451e9d-2c5b-421b-b219-09f1a2a5adae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.203854] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.204199] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.204415] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.204616] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.205127] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.205331] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.205572] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.205744] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.205917] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.206101] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.206296] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.212378] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5caaa811-3818-49a2-98b4-7335a2841abd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.228293] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 932.228293] env[63418]: value = "task-1245286" [ 932.228293] env[63418]: _type = "Task" [ 932.228293] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.236440] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245286, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.520838] env[63418]: DEBUG nova.compute.manager [req-aa5af117-0f09-4589-8275-ccdaa8c90e92 req-e708d479-3819-4a33-8ea8-cafb23cd8471 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Received event network-vif-plugged-1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 932.521153] env[63418]: DEBUG oslo_concurrency.lockutils [req-aa5af117-0f09-4589-8275-ccdaa8c90e92 req-e708d479-3819-4a33-8ea8-cafb23cd8471 service nova] Acquiring lock "26147513-3b25-4cc1-991a-34a724f73711-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.521581] env[63418]: DEBUG oslo_concurrency.lockutils [req-aa5af117-0f09-4589-8275-ccdaa8c90e92 req-e708d479-3819-4a33-8ea8-cafb23cd8471 service nova] Lock "26147513-3b25-4cc1-991a-34a724f73711-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.521889] env[63418]: DEBUG oslo_concurrency.lockutils [req-aa5af117-0f09-4589-8275-ccdaa8c90e92 req-e708d479-3819-4a33-8ea8-cafb23cd8471 service nova] Lock "26147513-3b25-4cc1-991a-34a724f73711-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.522122] env[63418]: DEBUG nova.compute.manager [req-aa5af117-0f09-4589-8275-ccdaa8c90e92 req-e708d479-3819-4a33-8ea8-cafb23cd8471 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] No waiting events found dispatching network-vif-plugged-1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 932.522397] env[63418]: WARNING nova.compute.manager [req-aa5af117-0f09-4589-8275-ccdaa8c90e92 req-e708d479-3819-4a33-8ea8-cafb23cd8471 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Received unexpected event network-vif-plugged-1d6d5d56-60f8-40db-b8f3-3026dfb63e33 for instance with vm_state building and task_state spawning. [ 932.535894] env[63418]: DEBUG nova.compute.utils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.537629] env[63418]: DEBUG nova.objects.instance [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lazy-loading 'numa_topology' on Instance uuid 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.539165] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 932.540029] env[63418]: DEBUG nova.network.neutron [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.554820] env[63418]: DEBUG oslo_vmware.api [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245283, 'name': PowerOnVM_Task, 'duration_secs': 1.01289} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.554820] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.554820] env[63418]: INFO nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Took 8.03 seconds to spawn the instance on the hypervisor. [ 932.554820] env[63418]: DEBUG nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 932.555440] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4896c0-661f-4d0d-8e21-77ddc334315c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.593201] env[63418]: DEBUG nova.policy [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea507bad11c3406d880ba47d08a047c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd445600834dd4c7e8022349ee993f3ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.620342] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245284, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572792} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.620662] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0f99b32a-0125-4df0-919c-e5456b4ae4a5/0f99b32a-0125-4df0-919c-e5456b4ae4a5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.620884] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.621151] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bff64d3f-7cca-4c77-8f0a-5a7a367cdd98 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.628367] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 932.628367] env[63418]: value = "task-1245287" [ 932.628367] env[63418]: _type = "Task" [ 932.628367] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.636848] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245287, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.738138] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245286, 'name': ReconfigVM_Task, 'duration_secs': 0.147842} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.738621] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 33 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.861610] env[63418]: DEBUG nova.network.neutron [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Successfully created port: fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.040303] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 933.046079] env[63418]: DEBUG nova.objects.base [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Object Instance<6465fb5c-7bc9-4197-b3fb-bad2c000a5f8> lazy-loaded attributes: resources,numa_topology {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 933.049237] env[63418]: DEBUG nova.network.neutron [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Successfully updated port: 1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.080597] env[63418]: INFO nova.compute.manager [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Unrescuing [ 933.080862] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.081022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquired lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.081195] env[63418]: DEBUG nova.network.neutron [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.094243] env[63418]: INFO nova.compute.manager [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Took 12.61 seconds to build instance. [ 933.097089] env[63418]: DEBUG nova.compute.manager [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Received event network-changed-1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 933.097282] env[63418]: DEBUG nova.compute.manager [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Refreshing instance network info cache due to event network-changed-1d6d5d56-60f8-40db-b8f3-3026dfb63e33. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 933.097737] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] Acquiring lock "refresh_cache-26147513-3b25-4cc1-991a-34a724f73711" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.098405] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] Acquired lock "refresh_cache-26147513-3b25-4cc1-991a-34a724f73711" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.098603] env[63418]: DEBUG nova.network.neutron [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Refreshing network info cache for port 1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.141886] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245287, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070162} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.142242] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.143346] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ae5c02-7db8-496d-aeb7-c345d1c13ebe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.168331] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 0f99b32a-0125-4df0-919c-e5456b4ae4a5/0f99b32a-0125-4df0-919c-e5456b4ae4a5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.171519] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90a66268-f290-4578-b87a-a6949ddf4dbc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.190899] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 933.190899] env[63418]: value = "task-1245288" [ 933.190899] env[63418]: _type = "Task" [ 933.190899] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.202809] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245288, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.245906] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.246148] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.246497] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.246497] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.246678] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.246719] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.246963] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.247163] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.247338] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.247506] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.247678] env[63418]: DEBUG nova.virt.hardware [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.252955] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Reconfiguring VM instance instance-00000040 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.256068] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63d15211-dc82-4a7d-b5fb-813e659f863d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.274120] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 933.274120] env[63418]: value = "task-1245289" [ 933.274120] env[63418]: _type = "Task" [ 933.274120] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.285599] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245289, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.391809] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d21c3a-bcc3-4ee2-a0cb-9100ff98b6e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.401990] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545c5c0d-cbde-4f02-9ca9-3c7ac7fa0a27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.439045] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abed1367-6f7f-45bd-bf22-d7a662a65fe2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.448212] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f133d96-529d-44eb-a4af-d956a37a8e08 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.464282] env[63418]: DEBUG nova.compute.provider_tree [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.551847] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-26147513-3b25-4cc1-991a-34a724f73711" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.600837] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b440c35-1cf7-4819-a058-64b98b54a481 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.145s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.636715] env[63418]: DEBUG nova.network.neutron [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.702831] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245288, 'name': ReconfigVM_Task, 'duration_secs': 0.336368} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.702831] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 0f99b32a-0125-4df0-919c-e5456b4ae4a5/0f99b32a-0125-4df0-919c-e5456b4ae4a5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.702831] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79557b41-bdc2-4de3-8c88-a8fdb473df51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.711198] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 933.711198] env[63418]: value = "task-1245290" [ 933.711198] env[63418]: _type = "Task" [ 933.711198] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.723211] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245290, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.728899] env[63418]: DEBUG nova.network.neutron [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.786591] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245289, 'name': ReconfigVM_Task, 'duration_secs': 0.189755} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.786591] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Reconfigured VM instance instance-00000040 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.787411] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a6d95f-0dbe-4f64-bcb9-ec51340d9142 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.811096] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a/7b0c70aa-e2bc-4131-97b4-4e53a378940a.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.814672] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1674adc-b7d2-4510-ae7d-04ee8919e3b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.834306] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 933.834306] env[63418]: value = "task-1245291" [ 933.834306] env[63418]: _type = "Task" [ 933.834306] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.849050] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245291, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.986766] env[63418]: ERROR nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [req-049e8335-c199-4502-963c-07b05e9f7244] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-049e8335-c199-4502-963c-07b05e9f7244"}]} [ 933.992391] env[63418]: DEBUG nova.network.neutron [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updating instance_info_cache with network_info: [{"id": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "address": "fa:16:3e:36:3a:6f", "network": {"id": "5d5aaf6a-993a-41a1-8156-8f564b8a123e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2113547273-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e074b295ca4d4e50bde88f6d37a7f98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b36c5ae6-c344-4bd1-8239-29128e2bbfbf", "external-id": "nsx-vlan-transportzone-214", "segmentation_id": 214, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef27d89f-f8", "ovs_interfaceid": "ef27d89f-f8f7-46cc-92e2-811aee3d2dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.004729] env[63418]: DEBUG nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 934.019645] env[63418]: DEBUG nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 934.019930] env[63418]: DEBUG nova.compute.provider_tree [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.031865] env[63418]: DEBUG nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 934.050923] env[63418]: DEBUG nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 934.054118] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 934.085134] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e9b1606f65523c1fd6301be4da67a965',container_format='bare',created_at=2024-10-10T13:44:26Z,direct_url=,disk_format='vmdk',id=3e5a3485-02fb-4c87-bf16-417423b1a071,min_disk=1,min_ram=0,name='tempest-test-snap-1175496023',owner='d445600834dd4c7e8022349ee993f3ef',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T13:44:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.085563] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.085853] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.086145] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.086400] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.086639] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.086940] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.087188] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.087439] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.087678] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.088074] env[63418]: DEBUG nova.virt.hardware [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.089487] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3164f051-c771-4db5-96f5-5abfad7ebfa9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.105950] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0206e1e2-8c53-49b3-bdc2-05209145de5f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.223245] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245290, 'name': Rename_Task, 'duration_secs': 0.142591} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.223565] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.223829] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2bd20a86-ba74-4885-a11c-324f831c385d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.231423] env[63418]: DEBUG oslo_concurrency.lockutils [req-9bd1c71a-face-4c1e-af08-2e438bc192f3 req-699aeb6a-75da-4252-8134-5193e3074011 service nova] Releasing lock "refresh_cache-26147513-3b25-4cc1-991a-34a724f73711" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.231779] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-26147513-3b25-4cc1-991a-34a724f73711" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.231977] env[63418]: DEBUG nova.network.neutron [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.234916] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 934.234916] env[63418]: value = "task-1245292" [ 934.234916] env[63418]: _type = "Task" [ 934.234916] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.246625] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.268372] env[63418]: DEBUG nova.compute.manager [req-991b9089-eb7d-442e-afd0-cb7f3c0d8b91 req-03941cb0-12f2-49ab-90b7-abaa521020da service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Received event network-vif-plugged-fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 934.268598] env[63418]: DEBUG oslo_concurrency.lockutils [req-991b9089-eb7d-442e-afd0-cb7f3c0d8b91 req-03941cb0-12f2-49ab-90b7-abaa521020da service nova] Acquiring lock "23af1d2e-f8ec-4c09-a070-9719b593f83a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.268852] env[63418]: DEBUG oslo_concurrency.lockutils [req-991b9089-eb7d-442e-afd0-cb7f3c0d8b91 req-03941cb0-12f2-49ab-90b7-abaa521020da service nova] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.269117] env[63418]: DEBUG oslo_concurrency.lockutils [req-991b9089-eb7d-442e-afd0-cb7f3c0d8b91 req-03941cb0-12f2-49ab-90b7-abaa521020da service nova] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.269367] env[63418]: DEBUG nova.compute.manager [req-991b9089-eb7d-442e-afd0-cb7f3c0d8b91 req-03941cb0-12f2-49ab-90b7-abaa521020da service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] No waiting events found dispatching network-vif-plugged-fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 934.269573] env[63418]: WARNING nova.compute.manager [req-991b9089-eb7d-442e-afd0-cb7f3c0d8b91 req-03941cb0-12f2-49ab-90b7-abaa521020da service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Received unexpected event network-vif-plugged-fe92a737-df95-4d64-ae0a-439e030ec74a for instance with vm_state building and task_state spawning. [ 934.288166] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78acdaf5-3ff9-4e5d-94b2-c748552017b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.296892] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fc8ff3-fd8d-4373-ad12-217ddbaa3315 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.330318] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a444003-3943-4f6f-9af9-5fec381b9e77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.342238] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c10470-f01f-4399-86c8-da3223a177e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.361166] env[63418]: DEBUG nova.compute.provider_tree [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.362660] env[63418]: DEBUG oslo_vmware.api [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245291, 'name': ReconfigVM_Task, 'duration_secs': 0.261782} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.363017] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a/7b0c70aa-e2bc-4131-97b4-4e53a378940a.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.363403] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 50 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 934.398391] env[63418]: DEBUG nova.network.neutron [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Successfully updated port: fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.495589] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Releasing lock "refresh_cache-24a97d0a-252b-4bbd-9dfc-57767ef8014d" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.496437] env[63418]: DEBUG nova.objects.instance [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lazy-loading 'flavor' on Instance uuid 24a97d0a-252b-4bbd-9dfc-57767ef8014d {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.747421] env[63418]: DEBUG oslo_vmware.api [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245292, 'name': PowerOnVM_Task, 'duration_secs': 0.469846} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.747725] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.747954] env[63418]: INFO nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Took 7.84 seconds to spawn the instance on the hypervisor. [ 934.748151] env[63418]: DEBUG nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 934.748945] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1869252e-f359-4a00-8e97-908b9d54e9c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.768420] env[63418]: DEBUG nova.network.neutron [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 934.868331] env[63418]: DEBUG nova.scheduler.client.report [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 934.876143] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106a0015-1dd2-470b-909c-7f8ecda5a4ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.897930] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6ddb63-064a-43e4-a11b-24308cded683 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.901221] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "refresh_cache-23af1d2e-f8ec-4c09-a070-9719b593f83a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.901369] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "refresh_cache-23af1d2e-f8ec-4c09-a070-9719b593f83a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.901517] env[63418]: DEBUG nova.network.neutron [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.922029] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 67 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 934.927195] env[63418]: DEBUG nova.network.neutron [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Updating instance_info_cache with network_info: [{"id": "1d6d5d56-60f8-40db-b8f3-3026dfb63e33", "address": "fa:16:3e:e0:e3:a3", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d6d5d56-60", "ovs_interfaceid": "1d6d5d56-60f8-40db-b8f3-3026dfb63e33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.002636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bec4cf0-f447-4d74-b474-a313cf1ec9bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.025583] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.025917] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6179d36b-f755-4a8b-b1df-1c9f6611cb3a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.035160] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 935.035160] env[63418]: value = "task-1245293" [ 935.035160] env[63418]: _type = "Task" [ 935.035160] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.044729] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245293, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.118632] env[63418]: DEBUG nova.compute.manager [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Received event network-changed-01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 935.118875] env[63418]: DEBUG nova.compute.manager [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Refreshing instance network info cache due to event network-changed-01df43f5-4069-46a0-b1e6-4baff131e04e. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 935.119043] env[63418]: DEBUG oslo_concurrency.lockutils [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] Acquiring lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.119200] env[63418]: DEBUG oslo_concurrency.lockutils [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] Acquired lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.119383] env[63418]: DEBUG nova.network.neutron [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Refreshing network info cache for port 01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.268910] env[63418]: INFO nova.compute.manager [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Took 14.54 seconds to build instance. [ 935.376447] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.344s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.378857] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.387s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.430095] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-26147513-3b25-4cc1-991a-34a724f73711" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.430498] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Instance network_info: |[{"id": "1d6d5d56-60f8-40db-b8f3-3026dfb63e33", "address": "fa:16:3e:e0:e3:a3", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d6d5d56-60", "ovs_interfaceid": "1d6d5d56-60f8-40db-b8f3-3026dfb63e33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 935.434828] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:e3:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d6d5d56-60f8-40db-b8f3-3026dfb63e33', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.444517] env[63418]: DEBUG oslo.service.loopingcall [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.445506] env[63418]: DEBUG nova.network.neutron [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.447575] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.448334] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c234fd36-1ad5-4e64-89dd-7f2d15a14992 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.474302] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.474302] env[63418]: value = "task-1245294" [ 935.474302] env[63418]: _type = "Task" [ 935.474302] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.486108] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245294, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.528847] env[63418]: DEBUG nova.network.neutron [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Port d0fa119b-63f3-4b54-8592-195a043ff0ee binding to destination host cpu-1 is already ACTIVE {{(pid=63418) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 935.552411] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245293, 'name': PowerOffVM_Task, 'duration_secs': 0.274218} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.557126] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.567196] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 935.568206] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10ac4139-3006-4802-be8c-ea4a849d4f30 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.600623] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 935.600623] env[63418]: value = "task-1245295" [ 935.600623] env[63418]: _type = "Task" [ 935.600623] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.614070] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.692175] env[63418]: DEBUG nova.network.neutron [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Updating instance_info_cache with network_info: [{"id": "fe92a737-df95-4d64-ae0a-439e030ec74a", "address": "fa:16:3e:76:23:e8", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe92a737-df", "ovs_interfaceid": "fe92a737-df95-4d64-ae0a-439e030ec74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.770638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3ff2b172-aeeb-45c3-b45e-4a12bf5e0e4f tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.054s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.886959] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e40ecce3-11db-41bc-8410-b7f91d0a569f tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 32.054s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.887833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 6.895s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.888034] env[63418]: INFO nova.compute.manager [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Unshelving [ 935.987576] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245294, 'name': CreateVM_Task, 'duration_secs': 0.386195} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.987736] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.988497] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.988731] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.989085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.989401] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2424b2e8-eee4-4a18-ab8b-ba9c1bf1c3aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.995369] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 935.995369] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52196f33-57a8-4d06-225e-13b709a96031" [ 935.995369] env[63418]: _type = "Task" [ 935.995369] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.010551] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52196f33-57a8-4d06-225e-13b709a96031, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.112863] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245295, 'name': ReconfigVM_Task, 'duration_secs': 0.315381} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.113954] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 936.114169] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.114925] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f7e6a6-75bb-43cf-8d26-0d43b4ff5ba4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.117333] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-368e3f7b-70c1-44ae-b31c-0e2fab812fe2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.124670] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1218b097-43e6-42d7-8bcd-54df3ae157e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.128933] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 936.128933] env[63418]: value = "task-1245296" [ 936.128933] env[63418]: _type = "Task" [ 936.128933] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.162413] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ac0e7f-42d5-4364-80c0-c20748fc2f39 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.168759] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245296, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.174328] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1cbb0c-68da-4882-b499-a1ad203cdc7c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.189737] env[63418]: DEBUG nova.compute.provider_tree [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.194701] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "refresh_cache-23af1d2e-f8ec-4c09-a070-9719b593f83a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.194975] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Instance network_info: |[{"id": "fe92a737-df95-4d64-ae0a-439e030ec74a", "address": "fa:16:3e:76:23:e8", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe92a737-df", "ovs_interfaceid": "fe92a737-df95-4d64-ae0a-439e030ec74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 936.195419] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:23:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe92a737-df95-4d64-ae0a-439e030ec74a', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.203527] env[63418]: DEBUG oslo.service.loopingcall [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.204359] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.204601] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfc4eb00-d55c-4874-9abe-06e40a4b9237 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.229677] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.229677] env[63418]: value = "task-1245297" [ 936.229677] env[63418]: _type = "Task" [ 936.229677] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.238775] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245297, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.509808] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52196f33-57a8-4d06-225e-13b709a96031, 'name': SearchDatastore_Task, 'duration_secs': 0.043071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.510159] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.510439] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.510740] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.510900] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.511127] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.511420] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5be69e6-3e9c-4fec-80cd-b8887e44e5a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.521440] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.521686] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.522574] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abd329e7-23c1-40b0-a56d-31c607b4f35a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.525909] env[63418]: DEBUG nova.network.neutron [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updated VIF entry in instance network info cache for port 01df43f5-4069-46a0-b1e6-4baff131e04e. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 936.526351] env[63418]: DEBUG nova.network.neutron [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [{"id": "01df43f5-4069-46a0-b1e6-4baff131e04e", "address": "fa:16:3e:69:39:54", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01df43f5-40", "ovs_interfaceid": "01df43f5-4069-46a0-b1e6-4baff131e04e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.532315] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 936.532315] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5263ffbe-00b4-2858-baf7-cace4c6f5094" [ 936.532315] env[63418]: _type = "Task" [ 936.532315] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.549156] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5263ffbe-00b4-2858-baf7-cace4c6f5094, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.557453] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.557595] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.557846] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.641977] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245296, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.661470] env[63418]: DEBUG nova.compute.manager [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Received event network-changed-fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 936.661754] env[63418]: DEBUG nova.compute.manager [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Refreshing instance network info cache due to event network-changed-fe92a737-df95-4d64-ae0a-439e030ec74a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 936.662352] env[63418]: DEBUG oslo_concurrency.lockutils [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] Acquiring lock "refresh_cache-23af1d2e-f8ec-4c09-a070-9719b593f83a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.662570] env[63418]: DEBUG oslo_concurrency.lockutils [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] Acquired lock "refresh_cache-23af1d2e-f8ec-4c09-a070-9719b593f83a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.662798] env[63418]: DEBUG nova.network.neutron [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Refreshing network info cache for port fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.695025] env[63418]: DEBUG nova.scheduler.client.report [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 936.741323] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245297, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.918439] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.029991] env[63418]: DEBUG oslo_concurrency.lockutils [req-cbfd7594-cce0-415b-b1f3-234c498db2ca req-52adb7ea-0f9a-4d56-853a-31a7f555caa2 service nova] Releasing lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.050017] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5263ffbe-00b4-2858-baf7-cace4c6f5094, 'name': SearchDatastore_Task, 'duration_secs': 0.013551} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.050017] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f5f6836-f424-43f7-86cf-8d642b9949f9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.058017] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 937.058017] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522b1388-d04a-2625-4fc4-fe72fb7c0ff7" [ 937.058017] env[63418]: _type = "Task" [ 937.058017] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.069176] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522b1388-d04a-2625-4fc4-fe72fb7c0ff7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.144026] env[63418]: DEBUG oslo_vmware.api [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245296, 'name': PowerOnVM_Task, 'duration_secs': 0.591059} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.145048] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.145048] env[63418]: DEBUG nova.compute.manager [None req-4abf43a3-8826-43aa-81d6-cc117ae19e27 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 937.146038] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b1d308-65b4-46b0-876e-5f3e324dea0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.172789] env[63418]: DEBUG nova.compute.manager [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 937.173021] env[63418]: DEBUG nova.compute.manager [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing instance network info cache due to event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 937.173636] env[63418]: DEBUG oslo_concurrency.lockutils [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.173828] env[63418]: DEBUG oslo_concurrency.lockutils [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.174039] env[63418]: DEBUG nova.network.neutron [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.243166] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245297, 'name': CreateVM_Task, 'duration_secs': 0.553009} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.243366] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 937.244113] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.244346] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.244715] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.244989] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0d02b1a-9d79-4d54-839d-2c4bf9f5b6e2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.250290] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 937.250290] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b92208-5973-4716-8f9b-66bc9649857f" [ 937.250290] env[63418]: _type = "Task" [ 937.250290] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.259440] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b92208-5973-4716-8f9b-66bc9649857f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.510626] env[63418]: DEBUG nova.network.neutron [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Updated VIF entry in instance network info cache for port fe92a737-df95-4d64-ae0a-439e030ec74a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.510803] env[63418]: DEBUG nova.network.neutron [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Updating instance_info_cache with network_info: [{"id": "fe92a737-df95-4d64-ae0a-439e030ec74a", "address": "fa:16:3e:76:23:e8", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe92a737-df", "ovs_interfaceid": "fe92a737-df95-4d64-ae0a-439e030ec74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.572175] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522b1388-d04a-2625-4fc4-fe72fb7c0ff7, 'name': SearchDatastore_Task, 'duration_secs': 0.013302} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.572594] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.572879] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 26147513-3b25-4cc1-991a-34a724f73711/26147513-3b25-4cc1-991a-34a724f73711.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.574369] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d3961f7-47f1-4229-8c9b-62da192c0431 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.582500] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 937.582500] env[63418]: value = "task-1245298" [ 937.582500] env[63418]: _type = "Task" [ 937.582500] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.594837] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245298, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.623359] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.624042] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.624042] env[63418]: DEBUG nova.network.neutron [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.706360] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.327s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.709814] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.700s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.710086] env[63418]: DEBUG nova.objects.instance [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid 14832e8e-3b4a-462e-84bb-d5a1e541d329 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.767681] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.768107] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Processing image 3e5a3485-02fb-4c87-bf16-417423b1a071 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.768432] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.768654] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.768951] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.774755] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d48078e7-d149-4706-bffa-118d2f72775d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.780625] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.780919] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.781965] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-134f58da-4ce5-4a59-a33d-101ed8480153 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.789464] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 937.789464] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526d9f19-7487-7dc0-9af7-c8f2d8851df9" [ 937.789464] env[63418]: _type = "Task" [ 937.789464] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.799170] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526d9f19-7487-7dc0-9af7-c8f2d8851df9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.014099] env[63418]: DEBUG oslo_concurrency.lockutils [req-0a277651-9f2d-4d9e-ad9c-f86999489b0f req-4b0716d1-623e-4ab8-bf09-979d0fd9f1c4 service nova] Releasing lock "refresh_cache-23af1d2e-f8ec-4c09-a070-9719b593f83a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.090017] env[63418]: DEBUG nova.network.neutron [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updated VIF entry in instance network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.090478] env[63418]: DEBUG nova.network.neutron [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.097324] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245298, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.320036] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 938.320679] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Fetch image to [datastore1] OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1/OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 938.320992] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Downloading stream optimized image 3e5a3485-02fb-4c87-bf16-417423b1a071 to [datastore1] OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1/OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1.vmdk on the data store datastore1 as vApp {{(pid=63418) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 938.321228] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Downloading image file data 3e5a3485-02fb-4c87-bf16-417423b1a071 to the ESX as VM named 'OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1' {{(pid=63418) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 938.328931] env[63418]: INFO nova.scheduler.client.report [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocation for migration 64ccaa38-9e84-42b5-8d37-6092250c1d58 [ 938.434619] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 938.434619] env[63418]: value = "resgroup-9" [ 938.434619] env[63418]: _type = "ResourcePool" [ 938.434619] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 938.435644] env[63418]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-df0afb70-1cca-478a-8b97-d2e455953050 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.465827] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease: (returnval){ [ 938.465827] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253a1e2-3ad6-bb9a-63da-105f33295dd1" [ 938.465827] env[63418]: _type = "HttpNfcLease" [ 938.465827] env[63418]: } obtained for vApp import into resource pool (val){ [ 938.465827] env[63418]: value = "resgroup-9" [ 938.465827] env[63418]: _type = "ResourcePool" [ 938.465827] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 938.466152] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the lease: (returnval){ [ 938.466152] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253a1e2-3ad6-bb9a-63da-105f33295dd1" [ 938.466152] env[63418]: _type = "HttpNfcLease" [ 938.466152] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 938.482109] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.482109] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253a1e2-3ad6-bb9a-63da-105f33295dd1" [ 938.482109] env[63418]: _type = "HttpNfcLease" [ 938.482109] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 938.516209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.516209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.516450] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.516531] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.517029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.520487] env[63418]: INFO nova.compute.manager [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Terminating instance [ 938.525223] env[63418]: DEBUG nova.network.neutron [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.595147] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245298, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.948192} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.598441] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 26147513-3b25-4cc1-991a-34a724f73711/26147513-3b25-4cc1-991a-34a724f73711.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.598441] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.598441] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f66cb0eb-008e-4d9a-8703-0415704ed52b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.598441] env[63418]: DEBUG oslo_concurrency.lockutils [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.599457] env[63418]: DEBUG nova.compute.manager [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-changed-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 938.599457] env[63418]: DEBUG nova.compute.manager [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing instance network info cache due to event network-changed-34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 938.599457] env[63418]: DEBUG oslo_concurrency.lockutils [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.599457] env[63418]: DEBUG oslo_concurrency.lockutils [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.599457] env[63418]: DEBUG nova.network.neutron [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.607421] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 938.607421] env[63418]: value = "task-1245300" [ 938.607421] env[63418]: _type = "Task" [ 938.607421] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.618668] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245300, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.620721] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4b1153-a9d0-4935-962c-18690c55bfd1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.629520] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc34e4e-c639-40e6-a2a4-ce8a5fb03a61 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.664351] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548e9dc1-3a5b-44e7-847b-1155f295d268 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.673473] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4e49b6-0f64-49f0-9d46-7fa2b27a270f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.690340] env[63418]: DEBUG nova.compute.provider_tree [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.769769] env[63418]: DEBUG nova.compute.manager [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-changed-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 938.770292] env[63418]: DEBUG nova.compute.manager [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing instance network info cache due to event network-changed-34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 938.770610] env[63418]: DEBUG oslo_concurrency.lockutils [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.848227] env[63418]: DEBUG oslo_concurrency.lockutils [None req-50138e7b-f4d5-4d40-8313-5e26d6eb876e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.376s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.979030] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.979030] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253a1e2-3ad6-bb9a-63da-105f33295dd1" [ 938.979030] env[63418]: _type = "HttpNfcLease" [ 938.979030] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 939.028682] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.036707] env[63418]: DEBUG nova.compute.manager [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 939.036707] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.036707] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9feb38f5-980c-4b80-814c-9f949288fa34 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.043782] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.044363] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 939.044681] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01f4a9e5-ebb7-4656-b200-c57706bc44ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.046336] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 939.046525] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 939.054050] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 939.054050] env[63418]: value = "task-1245301" [ 939.054050] env[63418]: _type = "Task" [ 939.054050] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.064084] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.118021] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245300, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086849} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.118380] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.119188] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf55f22-f500-44f5-a1d3-cc0d2f3cd666 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.143561] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 26147513-3b25-4cc1-991a-34a724f73711/26147513-3b25-4cc1-991a-34a724f73711.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.146218] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe7073ba-35d1-4b28-8955-abeb6e666d1a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.168574] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 939.168574] env[63418]: value = "task-1245302" [ 939.168574] env[63418]: _type = "Task" [ 939.168574] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.178397] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245302, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.193959] env[63418]: DEBUG nova.scheduler.client.report [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 939.428330] env[63418]: DEBUG nova.network.neutron [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updated VIF entry in instance network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.428857] env[63418]: DEBUG nova.network.neutron [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.476960] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.476960] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253a1e2-3ad6-bb9a-63da-105f33295dd1" [ 939.476960] env[63418]: _type = "HttpNfcLease" [ 939.476960] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 939.477347] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 939.477347] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5253a1e2-3ad6-bb9a-63da-105f33295dd1" [ 939.477347] env[63418]: _type = "HttpNfcLease" [ 939.477347] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 939.478415] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84000ce8-5366-422d-9ae3-5c34eb8e44ba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.489094] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52379f67-2744-3989-1e01-99f22aae4cdc/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 939.489225] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52379f67-2744-3989-1e01-99f22aae4cdc/disk-0.vmdk. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 939.576619] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-98aaafff-1697-4472-899b-8400628913a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.588257] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245301, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.589969] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064e9b5a-f5d8-47d3-8b09-d29f3e61fd75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.594867] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.595079] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquired lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.595280] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Forcefully refreshing network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 939.614226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d961338-ed04-4d4e-bb7e-a764c03bb661 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.623903] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 83 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.680230] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245302, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.699419] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.702750] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.450s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.702750] env[63418]: DEBUG nova.objects.instance [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'resources' on Instance uuid d76a008c-9bd9-420b-873d-4f7d7f25b8ca {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.719379] env[63418]: INFO nova.scheduler.client.report [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance 14832e8e-3b4a-462e-84bb-d5a1e541d329 [ 939.931954] env[63418]: DEBUG oslo_concurrency.lockutils [req-eb36c28d-c7f1-4075-b767-f0f8a37aef15 req-2a7b8358-a401-4449-bf75-48dca3203edb service nova] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.932621] env[63418]: DEBUG oslo_concurrency.lockutils [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.932830] env[63418]: DEBUG nova.network.neutron [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 939.957682] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.958697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.958697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.958697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.958697] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.961988] env[63418]: INFO nova.compute.manager [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Terminating instance [ 940.080440] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245301, 'name': PowerOffVM_Task, 'duration_secs': 0.53803} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.082827] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.083067] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 940.083603] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12148e99-a389-4d94-bd48-d2574132b3b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.131639] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e5a639-a8b7-430d-be43-7c2fd58d2107 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance '7b0c70aa-e2bc-4131-97b4-4e53a378940a' progress to 100 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.137310] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.167463] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.167922] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.168161] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Deleting the datastore file [datastore1] 24a97d0a-252b-4bbd-9dfc-57767ef8014d {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.168407] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ac3bd8f-c712-434b-8564-628908860c40 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.185381] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245302, 'name': ReconfigVM_Task, 'duration_secs': 0.765363} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.185740] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 940.185740] env[63418]: value = "task-1245304" [ 940.185740] env[63418]: _type = "Task" [ 940.185740] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.186790] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 26147513-3b25-4cc1-991a-34a724f73711/26147513-3b25-4cc1-991a-34a724f73711.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.189251] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e145526-e048-4f82-9a66-11d37e0205b2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.200282] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.201926] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 940.201926] env[63418]: value = "task-1245305" [ 940.201926] env[63418]: _type = "Task" [ 940.201926] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.221895] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245305, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.234398] env[63418]: DEBUG oslo_concurrency.lockutils [None req-315b399e-1fca-447a-ba3d-ea2eadd898fd tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "14832e8e-3b4a-462e-84bb-d5a1e541d329" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.726s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.328013] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 940.328289] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52379f67-2744-3989-1e01-99f22aae4cdc/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 940.329269] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eee58b-7f2e-43f2-90f8-bd1c9610d96a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.337214] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52379f67-2744-3989-1e01-99f22aae4cdc/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 940.337336] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52379f67-2744-3989-1e01-99f22aae4cdc/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 940.340363] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-45d56f5c-2696-4b56-a57e-a933e2d3250f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.455116] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.455400] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.458208] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054d8758-112d-45da-9bb2-4ffdff1f2a11 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.467866] env[63418]: DEBUG nova.compute.manager [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 940.467866] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 940.469347] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddfc66f-4a14-480c-b4e9-78f664c3255c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.472961] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08007b0-ad83-47d9-9277-5b223617d3b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.482938] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.510874] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e18b7040-84c9-4ff5-b89b-099b4c182e9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.516235] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdd9acb-cf09-4029-b9d1-79fc2f7a2a26 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.532805] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f1055a-abe3-407c-95c0-bccef174e97b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.539514] env[63418]: DEBUG oslo_vmware.rw_handles [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52379f67-2744-3989-1e01-99f22aae4cdc/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 940.540019] env[63418]: INFO nova.virt.vmwareapi.images [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Downloaded image file data 3e5a3485-02fb-4c87-bf16-417423b1a071 [ 940.540277] env[63418]: DEBUG oslo_vmware.api [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 940.540277] env[63418]: value = "task-1245306" [ 940.540277] env[63418]: _type = "Task" [ 940.540277] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.541628] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433cd263-30fd-4b93-b803-22a532e3f22a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.560008] env[63418]: DEBUG nova.compute.provider_tree [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.579432] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb2cfb75-bdd5-4607-8323-02c1f47e3ea0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.581946] env[63418]: DEBUG oslo_vmware.api [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245306, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.616308] env[63418]: INFO nova.virt.vmwareapi.images [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] The imported VM was unregistered [ 940.618890] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 940.619156] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating directory with path [datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.619444] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-176c1b63-7222-4d12-b627-6c43ad761600 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.632875] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created directory with path [datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.633170] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1/OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1.vmdk to [datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk. {{(pid=63418) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 940.633375] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e63d7877-1a60-4d7f-a5d1-a9a04cf2a6db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.645050] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 940.645050] env[63418]: value = "task-1245308" [ 940.645050] env[63418]: _type = "Task" [ 940.645050] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.657228] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.700891] env[63418]: DEBUG oslo_vmware.api [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256059} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.701166] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.701364] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.701582] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.701727] env[63418]: INFO nova.compute.manager [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Took 1.67 seconds to destroy the instance on the hypervisor. [ 940.701994] env[63418]: DEBUG oslo.service.loopingcall [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.702190] env[63418]: DEBUG nova.compute.manager [-] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 940.702306] env[63418]: DEBUG nova.network.neutron [-] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 940.714925] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245305, 'name': Rename_Task, 'duration_secs': 0.210787} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.715498] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.715780] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc0e6684-afda-416c-9dbb-a688f4630fc6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.723844] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 940.723844] env[63418]: value = "task-1245309" [ 940.723844] env[63418]: _type = "Task" [ 940.723844] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.732562] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.758949] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.783177] env[63418]: DEBUG nova.network.neutron [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updated VIF entry in instance network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 940.783613] env[63418]: DEBUG nova.network.neutron [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.963026] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 941.038426] env[63418]: DEBUG nova.compute.manager [req-a4c7df9a-c0b8-4923-ac43-7c6a5fe89880 req-0c816d59-17fb-4622-a0f0-74c26b90457b service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Received event network-vif-deleted-ef27d89f-f8f7-46cc-92e2-811aee3d2dea {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 941.038661] env[63418]: INFO nova.compute.manager [req-a4c7df9a-c0b8-4923-ac43-7c6a5fe89880 req-0c816d59-17fb-4622-a0f0-74c26b90457b service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Neutron deleted interface ef27d89f-f8f7-46cc-92e2-811aee3d2dea; detaching it from the instance and deleting it from the info cache [ 941.038932] env[63418]: DEBUG nova.network.neutron [req-a4c7df9a-c0b8-4923-ac43-7c6a5fe89880 req-0c816d59-17fb-4622-a0f0-74c26b90457b service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.056856] env[63418]: DEBUG oslo_vmware.api [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245306, 'name': PowerOffVM_Task, 'duration_secs': 0.222224} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.058079] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.059043] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.059043] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fce250a6-5d4b-4368-9add-ef4af9025adf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.063070] env[63418]: DEBUG nova.scheduler.client.report [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 941.110668] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.110786] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.159410] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.171776] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.172044] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.172165] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore2] 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.172424] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82c33546-b1cc-40c0-bd05-45ad92a35d55 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.180359] env[63418]: DEBUG oslo_vmware.api [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 941.180359] env[63418]: value = "task-1245311" [ 941.180359] env[63418]: _type = "Task" [ 941.180359] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.189272] env[63418]: DEBUG oslo_vmware.api [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.234955] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245309, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.262315] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Releasing lock "refresh_cache-d76a008c-9bd9-420b-873d-4f7d7f25b8ca" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.262585] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Updated the network info_cache for instance {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10349}} [ 941.262758] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.263023] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.263232] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.263432] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.263621] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.263770] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.263900] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 941.264057] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.288567] env[63418]: DEBUG oslo_concurrency.lockutils [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.288889] env[63418]: DEBUG nova.compute.manager [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 941.289087] env[63418]: DEBUG nova.compute.manager [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing instance network info cache due to event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 941.289314] env[63418]: DEBUG oslo_concurrency.lockutils [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.289463] env[63418]: DEBUG oslo_concurrency.lockutils [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.289629] env[63418]: DEBUG nova.network.neutron [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.481619] env[63418]: DEBUG nova.network.neutron [-] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.486980] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.541813] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a02c311c-646f-41e8-8634-2d2e210dc66a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.552634] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928cf490-bf63-41bd-8c91-b0783db6de77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.568044] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.570354] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.866s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.570634] env[63418]: DEBUG nova.objects.instance [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lazy-loading 'resources' on Instance uuid 4838d2b2-a187-48ae-a3b1-8ca24b961359 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.585027] env[63418]: DEBUG nova.compute.manager [req-a4c7df9a-c0b8-4923-ac43-7c6a5fe89880 req-0c816d59-17fb-4622-a0f0-74c26b90457b service nova] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Detach interface failed, port_id=ef27d89f-f8f7-46cc-92e2-811aee3d2dea, reason: Instance 24a97d0a-252b-4bbd-9dfc-57767ef8014d could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 941.589054] env[63418]: INFO nova.scheduler.client.report [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted allocations for instance d76a008c-9bd9-420b-873d-4f7d7f25b8ca [ 941.613992] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 941.665754] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.694646] env[63418]: DEBUG oslo_vmware.api [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245311, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.250989} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.694960] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.695177] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.695435] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.696115] env[63418]: INFO nova.compute.manager [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Took 1.23 seconds to destroy the instance on the hypervisor. [ 941.696115] env[63418]: DEBUG oslo.service.loopingcall [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.696115] env[63418]: DEBUG nova.compute.manager [-] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 941.696309] env[63418]: DEBUG nova.network.neutron [-] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.735112] env[63418]: DEBUG oslo_vmware.api [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245309, 'name': PowerOnVM_Task, 'duration_secs': 0.642735} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.735431] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.735638] env[63418]: INFO nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Took 10.05 seconds to spawn the instance on the hypervisor. [ 941.735828] env[63418]: DEBUG nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.736711] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f9fd5c-bc2f-4c38-bf5a-06011fb3ac10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.767795] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.984563] env[63418]: INFO nova.compute.manager [-] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Took 1.28 seconds to deallocate network for instance. [ 942.033971] env[63418]: DEBUG nova.network.neutron [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updated VIF entry in instance network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.034413] env[63418]: DEBUG nova.network.neutron [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.098320] env[63418]: DEBUG oslo_concurrency.lockutils [None req-72928027-c24e-480a-a4e1-72b913ce0d6c tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d76a008c-9bd9-420b-873d-4f7d7f25b8ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.666s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.135359] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.161653] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.258044] env[63418]: INFO nova.compute.manager [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Took 20.53 seconds to build instance. [ 942.337494] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a76a33b-1874-45b5-95fd-ccc772d0e18a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.346137] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcfcbf8-388e-4425-b1de-a0f1033c8dfb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.383314] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f21088-3aa5-4fe4-b8e0-5b04696831ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.392215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bd5fd7-1833-4d81-a8b1-f3a749d14d29 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.408061] env[63418]: DEBUG nova.compute.provider_tree [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.456420] env[63418]: DEBUG nova.network.neutron [-] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.491576] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.537492] env[63418]: DEBUG oslo_concurrency.lockutils [req-be807ef8-aa85-4a47-b276-b795314dd4d9 req-ea765203-aa3d-4a97-8a47-3af5a69c8d2b service nova] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.658826] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.759870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-28c93e70-9e1c-42e6-ae9a-713e46b0629f tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "26147513-3b25-4cc1-991a-34a724f73711" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.047s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.777849] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.778018] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.778225] env[63418]: DEBUG nova.compute.manager [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Going to confirm migration 2 {{(pid=63418) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 942.911765] env[63418]: DEBUG nova.scheduler.client.report [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 942.959081] env[63418]: INFO nova.compute.manager [-] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Took 1.26 seconds to deallocate network for instance. [ 943.066793] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d626cd78-5d0b-4d4f-9636-55eb12327af7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.072164] env[63418]: DEBUG nova.compute.manager [req-b05c45ef-3c3c-4f89-9a4d-57a06faa804d req-92146496-21f0-4aba-88a8-21367cd77664 service nova] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Received event network-vif-deleted-e4ff1dd4-2e2f-42cb-b882-19c984a2eb28 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 943.077905] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Suspending the VM {{(pid=63418) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 943.078239] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ace8f6e2-caa1-41bf-bbd5-bc90f28d7fa2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.086746] env[63418]: DEBUG oslo_vmware.api [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 943.086746] env[63418]: value = "task-1245312" [ 943.086746] env[63418]: _type = "Task" [ 943.086746] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.099794] env[63418]: DEBUG oslo_vmware.api [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245312, 'name': SuspendVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.158635] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.418937] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.421412] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.503s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.421652] env[63418]: DEBUG nova.objects.instance [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lazy-loading 'pci_requests' on Instance uuid 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.442666] env[63418]: INFO nova.scheduler.client.report [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Deleted allocations for instance 4838d2b2-a187-48ae-a3b1-8ca24b961359 [ 943.466180] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.591678] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.591945] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.592185] env[63418]: DEBUG nova.network.neutron [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.592438] env[63418]: DEBUG nova.objects.instance [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'info_cache' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.603267] env[63418]: DEBUG oslo_vmware.api [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245312, 'name': SuspendVM_Task} progress is 62%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.659240] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245308, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.640738} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.659514] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1/OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1.vmdk to [datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk. [ 943.659710] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Cleaning up location [datastore1] OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 943.659878] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_1853b96b-2713-42fa-83a8-7e0f26588be1 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.660151] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-785e78a7-9906-422f-aff2-993d83f97cef {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.668012] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 943.668012] env[63418]: value = "task-1245313" [ 943.668012] env[63418]: _type = "Task" [ 943.668012] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.689951] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.925387] env[63418]: DEBUG nova.objects.instance [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lazy-loading 'numa_topology' on Instance uuid 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.949962] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6ce18e5e-97e5-4ec4-9b55-35b855c65937 tempest-ImagesOneServerNegativeTestJSON-119962134 tempest-ImagesOneServerNegativeTestJSON-119962134-project-member] Lock "4838d2b2-a187-48ae-a3b1-8ca24b961359" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.650s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.099100] env[63418]: DEBUG oslo_vmware.api [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245312, 'name': SuspendVM_Task, 'duration_secs': 0.67676} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.099575] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Suspended the VM {{(pid=63418) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 944.099760] env[63418]: DEBUG nova.compute.manager [None req-45baef76-f3a3-444b-b141-18e379d352c0 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 944.100563] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3216dc3f-e3fb-4df4-b4e1-228f190ddfcd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.178045] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091112} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.178332] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.178634] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.178786] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk to [datastore1] 23af1d2e-f8ec-4c09-a070-9719b593f83a/23af1d2e-f8ec-4c09-a070-9719b593f83a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.179012] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-869c5d9f-2237-40c3-bee4-fb8940ea9d64 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.187022] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 944.187022] env[63418]: value = "task-1245314" [ 944.187022] env[63418]: _type = "Task" [ 944.187022] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.195222] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.529199] env[63418]: INFO nova.compute.claims [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.697314] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245314, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.910494] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.910711] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.136132] env[63418]: DEBUG nova.network.neutron [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.197478] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245314, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.412989] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 945.640526] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.640795] env[63418]: DEBUG nova.objects.instance [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'migration_context' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.697628] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245314, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.699468] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1e6ead-1828-4765-a9ec-e580654321bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.707836] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377344dc-0fcf-4dc9-a4a1-6ef2461d58e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.739943] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11622250-fe54-4085-aab4-c4db8c91a14e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.748821] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d1aea6-1066-4b99-9827-56c30f9ad58b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.766658] env[63418]: DEBUG nova.compute.provider_tree [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.842320] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "26147513-3b25-4cc1-991a-34a724f73711" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.842591] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "26147513-3b25-4cc1-991a-34a724f73711" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.842811] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "26147513-3b25-4cc1-991a-34a724f73711-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.843017] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "26147513-3b25-4cc1-991a-34a724f73711-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.843201] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "26147513-3b25-4cc1-991a-34a724f73711-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.845665] env[63418]: INFO nova.compute.manager [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Terminating instance [ 945.933388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.945242] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.945577] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.945984] env[63418]: DEBUG nova.objects.instance [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'flavor' on Instance uuid 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.143672] env[63418]: DEBUG nova.objects.base [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Object Instance<7b0c70aa-e2bc-4131-97b4-4e53a378940a> lazy-loaded attributes: info_cache,migration_context {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 946.144842] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcd84ce-93eb-497a-9d91-84fb143e6b79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.169969] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ec8180a-e539-460d-8e98-b5946d7838b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.177606] env[63418]: DEBUG oslo_vmware.api [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 946.177606] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c3b65d-a1e7-06c1-de22-c3e08bb62596" [ 946.177606] env[63418]: _type = "Task" [ 946.177606] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.188163] env[63418]: DEBUG oslo_vmware.api [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c3b65d-a1e7-06c1-de22-c3e08bb62596, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.197632] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245314, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.271677] env[63418]: DEBUG nova.scheduler.client.report [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 946.350036] env[63418]: DEBUG nova.compute.manager [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 946.350326] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.351359] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c91076f-2a0e-4125-826b-042c283736e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.361051] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.361341] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8cccfb3-cb62-4554-a4bf-15c04d1737ca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.440515] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.440515] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.440656] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleting the datastore file [datastore2] 26147513-3b25-4cc1-991a-34a724f73711 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.440963] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4926056d-d8b0-4880-87e4-bf1aa5ec593a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.451808] env[63418]: DEBUG oslo_vmware.api [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 946.451808] env[63418]: value = "task-1245316" [ 946.451808] env[63418]: _type = "Task" [ 946.451808] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.464557] env[63418]: DEBUG oslo_vmware.api [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.617043] env[63418]: DEBUG nova.objects.instance [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'pci_requests' on Instance uuid 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.690563] env[63418]: DEBUG oslo_vmware.api [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c3b65d-a1e7-06c1-de22-c3e08bb62596, 'name': SearchDatastore_Task, 'duration_secs': 0.046087} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.693762] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.699864] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245314, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.314754} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.700155] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3e5a3485-02fb-4c87-bf16-417423b1a071/3e5a3485-02fb-4c87-bf16-417423b1a071.vmdk to [datastore1] 23af1d2e-f8ec-4c09-a070-9719b593f83a/23af1d2e-f8ec-4c09-a070-9719b593f83a.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 946.700920] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d439f65a-957a-4797-80af-24481e216251 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.724937] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 23af1d2e-f8ec-4c09-a070-9719b593f83a/23af1d2e-f8ec-4c09-a070-9719b593f83a.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.725267] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d3cc7e6-96e4-4da8-b22a-e94f6a764bf6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.746112] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 946.746112] env[63418]: value = "task-1245317" [ 946.746112] env[63418]: _type = "Task" [ 946.746112] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.755945] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.777477] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.356s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.780197] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.293s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.781690] env[63418]: INFO nova.compute.claims [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.815872] env[63418]: INFO nova.network.neutron [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating port f786c95b-5214-454d-86f8-6d922f0482d8 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 946.963694] env[63418]: DEBUG oslo_vmware.api [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171547} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.964031] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.964222] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.964434] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.964634] env[63418]: INFO nova.compute.manager [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Took 0.61 seconds to destroy the instance on the hypervisor. [ 946.964881] env[63418]: DEBUG oslo.service.loopingcall [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.965096] env[63418]: DEBUG nova.compute.manager [-] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 946.965192] env[63418]: DEBUG nova.network.neutron [-] [instance: 26147513-3b25-4cc1-991a-34a724f73711] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 947.120033] env[63418]: DEBUG nova.objects.base [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Object Instance<039f4cf6-137b-4356-a1b8-dd47f8f0d6f5> lazy-loaded attributes: flavor,pci_requests {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 947.120296] env[63418]: DEBUG nova.network.neutron [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.223800] env[63418]: DEBUG nova.policy [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.256463] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245317, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.375744] env[63418]: DEBUG nova.compute.manager [req-4cf10364-4903-4944-be39-4382136a0cbd req-82880bcc-4aa3-43f9-9e10-e46d543d0b3a service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Received event network-vif-deleted-1d6d5d56-60f8-40db-b8f3-3026dfb63e33 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 947.375952] env[63418]: INFO nova.compute.manager [req-4cf10364-4903-4944-be39-4382136a0cbd req-82880bcc-4aa3-43f9-9e10-e46d543d0b3a service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Neutron deleted interface 1d6d5d56-60f8-40db-b8f3-3026dfb63e33; detaching it from the instance and deleting it from the info cache [ 947.376136] env[63418]: DEBUG nova.network.neutron [req-4cf10364-4903-4944-be39-4382136a0cbd req-82880bcc-4aa3-43f9-9e10-e46d543d0b3a service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.756469] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245317, 'name': ReconfigVM_Task, 'duration_secs': 0.520632} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.756777] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 23af1d2e-f8ec-4c09-a070-9719b593f83a/23af1d2e-f8ec-4c09-a070-9719b593f83a.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.757432] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fe6d86d-0643-4aca-8938-a81b5d90d8ff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.764872] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 947.764872] env[63418]: value = "task-1245318" [ 947.764872] env[63418]: _type = "Task" [ 947.764872] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.773216] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245318, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.793122] env[63418]: DEBUG nova.network.neutron [-] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.878985] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef2a757f-d4f7-49f7-bed1-268274491f40 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.888587] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fec6d6-5561-4574-b03b-0336dd4e7af4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.921986] env[63418]: DEBUG nova.compute.manager [req-4cf10364-4903-4944-be39-4382136a0cbd req-82880bcc-4aa3-43f9-9e10-e46d543d0b3a service nova] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Detach interface failed, port_id=1d6d5d56-60f8-40db-b8f3-3026dfb63e33, reason: Instance 26147513-3b25-4cc1-991a-34a724f73711 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 948.015153] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bd0532-3db4-4f82-873b-dd809a18c3ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.022941] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2209655e-e63a-44e2-b7ee-dc38a0548a6e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.051666] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c43efd2-174c-49ec-91a8-f5bd6c9aacb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.059131] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80a4fa4-f4d3-4443-8f10-f845168305fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.073036] env[63418]: DEBUG nova.compute.provider_tree [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.212984] env[63418]: DEBUG nova.compute.manager [req-9e4fa147-1be2-474b-b246-781bb25e4d46 req-fec36480-e80b-4dca-9361-9a1f9ed37897 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-vif-plugged-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 948.214596] env[63418]: DEBUG oslo_concurrency.lockutils [req-9e4fa147-1be2-474b-b246-781bb25e4d46 req-fec36480-e80b-4dca-9361-9a1f9ed37897 service nova] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.214871] env[63418]: DEBUG oslo_concurrency.lockutils [req-9e4fa147-1be2-474b-b246-781bb25e4d46 req-fec36480-e80b-4dca-9361-9a1f9ed37897 service nova] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.214991] env[63418]: DEBUG oslo_concurrency.lockutils [req-9e4fa147-1be2-474b-b246-781bb25e4d46 req-fec36480-e80b-4dca-9361-9a1f9ed37897 service nova] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.215176] env[63418]: DEBUG nova.compute.manager [req-9e4fa147-1be2-474b-b246-781bb25e4d46 req-fec36480-e80b-4dca-9361-9a1f9ed37897 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] No waiting events found dispatching network-vif-plugged-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 948.215368] env[63418]: WARNING nova.compute.manager [req-9e4fa147-1be2-474b-b246-781bb25e4d46 req-fec36480-e80b-4dca-9361-9a1f9ed37897 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received unexpected event network-vif-plugged-f786c95b-5214-454d-86f8-6d922f0482d8 for instance with vm_state shelved_offloaded and task_state spawning. [ 948.274623] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245318, 'name': Rename_Task, 'duration_secs': 0.147087} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.274865] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.275134] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47e86956-e492-4b8f-ae8c-a5fb67a8174e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.281619] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 948.281619] env[63418]: value = "task-1245319" [ 948.281619] env[63418]: _type = "Task" [ 948.281619] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.295290] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.295460] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.295655] env[63418]: DEBUG nova.network.neutron [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.297271] env[63418]: INFO nova.compute.manager [-] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Took 1.33 seconds to deallocate network for instance. [ 948.576112] env[63418]: DEBUG nova.scheduler.client.report [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 948.721576] env[63418]: DEBUG nova.network.neutron [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Successfully updated port: f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.793322] env[63418]: DEBUG oslo_vmware.api [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245319, 'name': PowerOnVM_Task, 'duration_secs': 0.505759} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.793727] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.793940] env[63418]: INFO nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Took 14.74 seconds to spawn the instance on the hypervisor. [ 948.794144] env[63418]: DEBUG nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 948.795299] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe4f1f0-8915-4f07-af8f-80411c4324c0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.806945] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.999008] env[63418]: DEBUG nova.network.neutron [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.081877] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.082557] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 949.085057] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.318s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.085275] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.085440] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 949.085727] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.951s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.087201] env[63418]: INFO nova.compute.claims [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.090389] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72057c28-d8ff-458f-8f4c-81540bacf140 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.099384] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac23c43-a781-4f7c-86cc-d07109cf642f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.113876] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d235c9-c315-43fc-af2e-cc398e67c094 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.122042] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b7e839-09e7-4f56-a6b1-94aff5b755c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.152852] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180680MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 949.153034] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.225477] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.225477] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.225477] env[63418]: DEBUG nova.network.neutron [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.010736] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.013611] env[63418]: DEBUG nova.compute.utils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.018861] env[63418]: DEBUG nova.compute.manager [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-vif-plugged-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 950.019077] env[63418]: DEBUG oslo_concurrency.lockutils [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.019282] env[63418]: DEBUG oslo_concurrency.lockutils [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.019444] env[63418]: DEBUG oslo_concurrency.lockutils [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.019611] env[63418]: DEBUG nova.compute.manager [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] No waiting events found dispatching network-vif-plugged-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 950.019768] env[63418]: WARNING nova.compute.manager [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received unexpected event network-vif-plugged-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 for instance with vm_state active and task_state None. [ 950.020123] env[63418]: DEBUG nova.compute.manager [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-changed-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 950.020123] env[63418]: DEBUG nova.compute.manager [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing instance network info cache due to event network-changed-f1f171f0-8f67-4b5b-a9c8-91014404b4e2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 950.020251] env[63418]: DEBUG oslo_concurrency.lockutils [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.025172] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 950.025336] env[63418]: DEBUG nova.network.neutron [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 950.031847] env[63418]: INFO nova.compute.manager [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Took 28.29 seconds to build instance. [ 950.054809] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='4e284bcd99d4231858ea26e287844f93',container_format='bare',created_at=2024-10-10T13:44:26Z,direct_url=,disk_format='vmdk',id=69a22df6-0b8e-4d81-ab99-3f97aaaed86e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-685955436-shelved',owner='0462b212fa4449c2a6f98cec2f186f51',properties=ImageMetaProps,protected=,size=31664128,status='active',tags=,updated_at=2024-10-10T13:44:42Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.055094] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.055261] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.055498] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.055605] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.055750] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.055959] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.056138] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.056575] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.056575] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.056762] env[63418]: DEBUG nova.virt.hardware [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.057844] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ee2bc1-475c-41cc-8543-76b0d1c06081 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.060990] env[63418]: WARNING nova.network.neutron [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] f11c9d94-8e8f-4c7d-b518-f8958080b8fa already exists in list: networks containing: ['f11c9d94-8e8f-4c7d-b518-f8958080b8fa']. ignoring it [ 950.069094] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1f8ce1-6038-4fd7-bca5-689a7211d717 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.082860] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:c4:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f786c95b-5214-454d-86f8-6d922f0482d8', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.090177] env[63418]: DEBUG oslo.service.loopingcall [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.090410] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 950.090625] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2dfaaaa-b6cb-4845-a215-d9a2425c1139 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.105262] env[63418]: DEBUG nova.policy [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ef8f62bd75a4d109a2e1881408754cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11fffb9badce4abeadce3ab70dff7d58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 950.116083] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.116083] env[63418]: value = "task-1245320" [ 950.116083] env[63418]: _type = "Task" [ 950.116083] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.123740] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245320, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.317373] env[63418]: DEBUG nova.compute.manager [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-changed-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 950.317612] env[63418]: DEBUG nova.compute.manager [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Refreshing instance network info cache due to event network-changed-f786c95b-5214-454d-86f8-6d922f0482d8. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 950.317840] env[63418]: DEBUG oslo_concurrency.lockutils [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.318228] env[63418]: DEBUG oslo_concurrency.lockutils [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.318228] env[63418]: DEBUG nova.network.neutron [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Refreshing network info cache for port f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.370544] env[63418]: DEBUG nova.network.neutron [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "address": "fa:16:3e:be:8e:cc", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1f171f0-8f", "ovs_interfaceid": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.416193] env[63418]: DEBUG nova.network.neutron [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Successfully created port: 7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.526496] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 950.534855] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bc55e22b-2b63-4ba0-b64a-870b25d319ac tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.815s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.538349] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "23af1d2e-f8ec-4c09-a070-9719b593f83a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.538544] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.538730] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "23af1d2e-f8ec-4c09-a070-9719b593f83a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.538930] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.539117] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.541587] env[63418]: INFO nova.compute.manager [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Terminating instance [ 950.626707] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245320, 'name': CreateVM_Task, 'duration_secs': 0.308749} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.626707] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.626962] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.627153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.627530] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.627785] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a785d7e0-1ce7-452f-bb6b-cc4233406646 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.633053] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 950.633053] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5241f8f3-6aa8-bb2c-51c8-465235b02c17" [ 950.633053] env[63418]: _type = "Task" [ 950.633053] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.641554] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5241f8f3-6aa8-bb2c-51c8-465235b02c17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.798140] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a89c33-db33-46b6-afbf-0995a1efb830 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.806322] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b479370-ed82-4d4d-a647-625116f6eec6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.842024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0955e6dc-2ecf-4583-8d55-9517d0f46224 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.850156] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4eb1ca5-8cfc-4c9b-acc7-40cc38973612 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.863623] env[63418]: DEBUG nova.compute.provider_tree [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.873053] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.873671] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.873836] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.874423] env[63418]: DEBUG oslo_concurrency.lockutils [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.874632] env[63418]: DEBUG nova.network.neutron [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing network info cache for port f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.876701] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc64017-1cf2-42b5-9cb9-1e398e1e75dc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.895700] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.895968] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.896148] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.896332] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.896479] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.897578] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.897578] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.897578] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.897578] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.897578] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.897578] env[63418]: DEBUG nova.virt.hardware [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.903973] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Reconfiguring VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 950.904451] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33ef3120-50ac-41b2-9903-0194d000d160 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.925353] env[63418]: DEBUG oslo_vmware.api [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 950.925353] env[63418]: value = "task-1245321" [ 950.925353] env[63418]: _type = "Task" [ 950.925353] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.934272] env[63418]: DEBUG oslo_vmware.api [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245321, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.047876] env[63418]: DEBUG nova.compute.manager [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 951.048110] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 951.048953] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5435f868-e3e2-4494-926d-a9372812d6a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.057766] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.058057] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f52706a4-dc67-421b-8916-41c68a8ffa0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.064479] env[63418]: DEBUG nova.network.neutron [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updated VIF entry in instance network info cache for port f786c95b-5214-454d-86f8-6d922f0482d8. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.064840] env[63418]: DEBUG nova.network.neutron [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.067236] env[63418]: DEBUG oslo_vmware.api [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 951.067236] env[63418]: value = "task-1245322" [ 951.067236] env[63418]: _type = "Task" [ 951.067236] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.077595] env[63418]: DEBUG oslo_vmware.api [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.144733] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.144991] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Processing image 69a22df6-0b8e-4d81-ab99-3f97aaaed86e {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.145251] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.145440] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.145623] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.145884] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b7ddd1b-f1da-4dc0-922a-7c1bda1ff020 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.156206] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.156389] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 951.157148] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eae60ca-8433-4fd9-b392-15aa761243ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.162679] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 951.162679] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528ebd64-d4f6-bd65-9f46-aed823d81682" [ 951.162679] env[63418]: _type = "Task" [ 951.162679] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.170628] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528ebd64-d4f6-bd65-9f46-aed823d81682, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.366783] env[63418]: DEBUG nova.scheduler.client.report [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 951.437626] env[63418]: DEBUG oslo_vmware.api [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245321, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.538977] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 951.564701] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.564990] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.565169] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.565383] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.565519] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.565665] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.565880] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.566052] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.566231] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.566400] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.566579] env[63418]: DEBUG nova.virt.hardware [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.567584] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64845670-44f0-4d1d-947f-ffa165dc6ebc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.570316] env[63418]: DEBUG oslo_concurrency.lockutils [req-28b1345c-a640-4c4d-9a21-a3eb1071e152 req-70bb73a0-3437-4a06-ae44-87dd96dd4be5 service nova] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.579232] env[63418]: DEBUG oslo_vmware.api [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245322, 'name': PowerOffVM_Task, 'duration_secs': 0.189278} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.581195] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.581375] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.582151] env[63418]: DEBUG nova.network.neutron [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updated VIF entry in instance network info cache for port f1f171f0-8f67-4b5b-a9c8-91014404b4e2. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.582534] env[63418]: DEBUG nova.network.neutron [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "address": "fa:16:3e:be:8e:cc", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1f171f0-8f", "ovs_interfaceid": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.583680] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-221c5955-8b51-42bb-8b8f-254b83a3314b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.586696] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c5db72-06b7-4c54-aa4b-dd208bdafc3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.658090] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.658476] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.658568] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleting the datastore file [datastore1] 23af1d2e-f8ec-4c09-a070-9719b593f83a {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.658809] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3c35678-346a-4631-b416-4d8285be633f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.668220] env[63418]: DEBUG oslo_vmware.api [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 951.668220] env[63418]: value = "task-1245324" [ 951.668220] env[63418]: _type = "Task" [ 951.668220] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.674991] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 951.675260] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Fetch image to [datastore1] OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea/OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 951.675452] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Downloading stream optimized image 69a22df6-0b8e-4d81-ab99-3f97aaaed86e to [datastore1] OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea/OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea.vmdk on the data store datastore1 as vApp {{(pid=63418) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 951.675626] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Downloading image file data 69a22df6-0b8e-4d81-ab99-3f97aaaed86e to the ESX as VM named 'OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea' {{(pid=63418) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 951.681681] env[63418]: DEBUG oslo_vmware.api [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.751206] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 951.751206] env[63418]: value = "resgroup-9" [ 951.751206] env[63418]: _type = "ResourcePool" [ 951.751206] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 951.751556] env[63418]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-13dbed83-40bc-4c68-bcfb-c17c856ca453 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.772611] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lease: (returnval){ [ 951.772611] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529b7527-20bb-4aa1-4680-3b1dda1d91bc" [ 951.772611] env[63418]: _type = "HttpNfcLease" [ 951.772611] env[63418]: } obtained for vApp import into resource pool (val){ [ 951.772611] env[63418]: value = "resgroup-9" [ 951.772611] env[63418]: _type = "ResourcePool" [ 951.772611] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 951.773074] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the lease: (returnval){ [ 951.773074] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529b7527-20bb-4aa1-4680-3b1dda1d91bc" [ 951.773074] env[63418]: _type = "HttpNfcLease" [ 951.773074] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 951.779741] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 951.779741] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529b7527-20bb-4aa1-4680-3b1dda1d91bc" [ 951.779741] env[63418]: _type = "HttpNfcLease" [ 951.779741] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 951.871757] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.786s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.872428] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 951.876031] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.385s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.876288] env[63418]: DEBUG nova.objects.instance [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lazy-loading 'resources' on Instance uuid 24a97d0a-252b-4bbd-9dfc-57767ef8014d {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.937305] env[63418]: DEBUG oslo_vmware.api [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245321, 'name': ReconfigVM_Task, 'duration_secs': 0.61253} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.938060] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.938250] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Reconfigured VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 952.091508] env[63418]: DEBUG oslo_concurrency.lockutils [req-8d98bd79-b912-42b5-82fb-46d25251e6e4 req-15743d47-33e6-4ea5-a4d3-f4b15f1a17f1 service nova] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.098087] env[63418]: DEBUG nova.network.neutron [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Successfully updated port: 7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.178831] env[63418]: DEBUG oslo_vmware.api [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139238} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.179140] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.179349] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 952.179524] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 952.179725] env[63418]: INFO nova.compute.manager [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 952.179983] env[63418]: DEBUG oslo.service.loopingcall [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.180199] env[63418]: DEBUG nova.compute.manager [-] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 952.180317] env[63418]: DEBUG nova.network.neutron [-] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 952.282023] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 952.282023] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529b7527-20bb-4aa1-4680-3b1dda1d91bc" [ 952.282023] env[63418]: _type = "HttpNfcLease" [ 952.282023] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 952.282351] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 952.282351] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529b7527-20bb-4aa1-4680-3b1dda1d91bc" [ 952.282351] env[63418]: _type = "HttpNfcLease" [ 952.282351] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 952.283109] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7123435-87a1-4035-9ed8-a23024aa1599 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.291874] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523cab77-4291-95d7-b841-05c8e9a0958d/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 952.291874] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating HTTP connection to write to file with size = 31664128 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523cab77-4291-95d7-b841-05c8e9a0958d/disk-0.vmdk. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 952.355960] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-deca6e94-72e0-41be-8690-b84879652f5a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.381016] env[63418]: DEBUG nova.compute.utils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.390993] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 952.391192] env[63418]: DEBUG nova.network.neutron [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 952.445195] env[63418]: DEBUG oslo_concurrency.lockutils [None req-09a42728-7440-484b-86d0-76ad853e41dc tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.499s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.454564] env[63418]: DEBUG nova.policy [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 952.465658] env[63418]: DEBUG nova.compute.manager [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Received event network-vif-plugged-7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 952.467052] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] Acquiring lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.467052] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.467438] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.467438] env[63418]: DEBUG nova.compute.manager [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] No waiting events found dispatching network-vif-plugged-7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 952.467583] env[63418]: WARNING nova.compute.manager [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Received unexpected event network-vif-plugged-7ab35b58-105e-4976-b7f7-05e0c77de3fa for instance with vm_state building and task_state spawning. [ 952.467745] env[63418]: DEBUG nova.compute.manager [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Received event network-changed-7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 952.467898] env[63418]: DEBUG nova.compute.manager [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Refreshing instance network info cache due to event network-changed-7ab35b58-105e-4976-b7f7-05e0c77de3fa. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 952.468991] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] Acquiring lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.468991] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] Acquired lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.468991] env[63418]: DEBUG nova.network.neutron [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Refreshing network info cache for port 7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.603141] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.653761] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17578c17-7990-48cd-8acd-ba4a0f8a645d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.667135] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8a4e7d-4caa-4dbd-bfa5-f4520a3e4f3a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.707215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb477de-4e1e-49ca-9c1b-6d009217971c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.720794] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9205768-a91c-4081-9c4b-cd0695805332 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.739923] env[63418]: DEBUG nova.compute.provider_tree [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.741905] env[63418]: DEBUG nova.network.neutron [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Successfully created port: bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.751247] env[63418]: DEBUG nova.scheduler.client.report [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 952.892189] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 952.944035] env[63418]: DEBUG nova.network.neutron [-] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.019724] env[63418]: DEBUG nova.network.neutron [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 953.132028] env[63418]: DEBUG nova.network.neutron [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.256995] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.381s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.259907] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.794s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.259907] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.261849] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.329s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.264061] env[63418]: INFO nova.compute.claims [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.280460] env[63418]: INFO nova.scheduler.client.report [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Deleted allocations for instance 24a97d0a-252b-4bbd-9dfc-57767ef8014d [ 953.289079] env[63418]: INFO nova.scheduler.client.report [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocations for instance 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5 [ 953.450033] env[63418]: INFO nova.compute.manager [-] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Took 1.27 seconds to deallocate network for instance. [ 953.573421] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 953.573764] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523cab77-4291-95d7-b841-05c8e9a0958d/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 953.574926] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c868b4-550d-4d69-8e90-0b8bcefcb12f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.583908] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523cab77-4291-95d7-b841-05c8e9a0958d/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 953.584109] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523cab77-4291-95d7-b841-05c8e9a0958d/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 953.584399] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-2184b34b-9700-464f-a853-eaab28f4506d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.635859] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7773177-1f2a-4749-a9af-754d17fbd74c req-c5e9752a-3fa2-447c-9dd9-207638e7f3e1 service nova] Releasing lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.637219] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.637219] env[63418]: DEBUG nova.network.neutron [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.776047] env[63418]: DEBUG oslo_vmware.rw_handles [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523cab77-4291-95d7-b841-05c8e9a0958d/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 953.776311] env[63418]: INFO nova.virt.vmwareapi.images [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Downloaded image file data 69a22df6-0b8e-4d81-ab99-3f97aaaed86e [ 953.777571] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5867b6f-25f9-4974-945d-6f5288a086b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.796359] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d7f949ba-8273-47bf-b141-1a0bf89a39b8 tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "24a97d0a-252b-4bbd-9dfc-57767ef8014d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.280s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.799896] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3b0f8fc1-589f-447e-b069-b54a87567e91 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.842s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.801547] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0379e2d-d0ce-4f63-a6ba-0e91e414a9a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.831123] env[63418]: INFO nova.virt.vmwareapi.images [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] The imported VM was unregistered [ 953.834028] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 953.834281] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Creating directory with path [datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.834471] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8aee25b6-4969-4f97-9e03-cfa7e4b31701 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.870468] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Created directory with path [datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.870468] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea/OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea.vmdk to [datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk. {{(pid=63418) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 953.870655] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-941f32df-7dd0-4030-8a2e-51b4304cd28c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.879636] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 953.879636] env[63418]: value = "task-1245327" [ 953.879636] env[63418]: _type = "Task" [ 953.879636] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.888879] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.902556] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 953.928305] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.928574] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.928743] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.928970] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.929148] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.929302] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.929515] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.929680] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.929853] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.930035] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.930247] env[63418]: DEBUG nova.virt.hardware [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.931129] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663e1e76-fb00-4c40-a80f-06c5cb2471b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.940363] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f8c1e6-44d6-465a-9f7f-e919de1b8274 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.957610] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.163629] env[63418]: DEBUG nova.compute.manager [req-4e6fc137-5e2d-44c6-bd28-13e85f5f56ea req-6f41cfb3-6203-49d9-aa7e-33082ba2776c service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Received event network-vif-plugged-bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 954.163940] env[63418]: DEBUG oslo_concurrency.lockutils [req-4e6fc137-5e2d-44c6-bd28-13e85f5f56ea req-6f41cfb3-6203-49d9-aa7e-33082ba2776c service nova] Acquiring lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.164301] env[63418]: DEBUG oslo_concurrency.lockutils [req-4e6fc137-5e2d-44c6-bd28-13e85f5f56ea req-6f41cfb3-6203-49d9-aa7e-33082ba2776c service nova] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.164361] env[63418]: DEBUG oslo_concurrency.lockutils [req-4e6fc137-5e2d-44c6-bd28-13e85f5f56ea req-6f41cfb3-6203-49d9-aa7e-33082ba2776c service nova] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.164589] env[63418]: DEBUG nova.compute.manager [req-4e6fc137-5e2d-44c6-bd28-13e85f5f56ea req-6f41cfb3-6203-49d9-aa7e-33082ba2776c service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] No waiting events found dispatching network-vif-plugged-bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 954.164801] env[63418]: WARNING nova.compute.manager [req-4e6fc137-5e2d-44c6-bd28-13e85f5f56ea req-6f41cfb3-6203-49d9-aa7e-33082ba2776c service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Received unexpected event network-vif-plugged-bae1fde8-1485-472f-9c86-34b47e93ee43 for instance with vm_state building and task_state spawning. [ 954.167345] env[63418]: DEBUG nova.network.neutron [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.170914] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.171153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.252221] env[63418]: DEBUG nova.network.neutron [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Successfully updated port: bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 954.362127] env[63418]: DEBUG nova.network.neutron [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [{"id": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "address": "fa:16:3e:43:d6:a0", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ab35b58-10", "ovs_interfaceid": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.393272] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.495022] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe45847-0fd2-4284-a330-1b07b2685d68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.504181] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68d7ed0-107a-48bf-9041-3cfa8e2fff5a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.539081] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e921a0-db4d-4cd4-b092-113de56b958e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.547410] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c4b673-5b0d-4092-814a-d12204228ae3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.563609] env[63418]: DEBUG nova.compute.provider_tree [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.573931] env[63418]: DEBUG nova.compute.manager [req-ae40635e-fa52-4247-8c7e-7c85b66ec73a req-f2c81c48-d09b-41fd-9870-27c7f4ecf751 service nova] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Received event network-vif-deleted-fe92a737-df95-4d64-ae0a-439e030ec74a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 954.675032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.675032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.675545] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247858e6-e342-465e-b187-9d2abc362689 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.696384] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df961b9-a6dc-4c64-952b-a41e6a207cd2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.723829] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Reconfiguring VM to detach interface {{(pid=63418) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 954.724184] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a20a6a28-de72-45b8-8502-0395ba3cf37d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.746704] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 954.746704] env[63418]: value = "task-1245328" [ 954.746704] env[63418]: _type = "Task" [ 954.746704] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.756259] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-b0887bcd-7ba7-4c0e-8d50-886e27d37649" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.756259] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-b0887bcd-7ba7-4c0e-8d50-886e27d37649" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.756259] env[63418]: DEBUG nova.network.neutron [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.758055] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.873085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.873085] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Instance network_info: |[{"id": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "address": "fa:16:3e:43:d6:a0", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ab35b58-10", "ovs_interfaceid": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 954.873415] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.873688] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.873870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.874079] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.874250] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.876747] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:d6:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ab35b58-105e-4976-b7f7-05e0c77de3fa', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.887539] env[63418]: DEBUG oslo.service.loopingcall [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.888292] env[63418]: INFO nova.compute.manager [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Terminating instance [ 954.890827] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.895967] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73f5bca1-fe7d-4d87-b5d8-e96ef8692e21 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.911666] env[63418]: DEBUG nova.compute.manager [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 954.911899] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.913067] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532e9e48-10db-47be-9668-e8b622c894af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.922517] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.924913] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.926250] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16e6fb34-4190-4618-9b53-24e53cefa839 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.927857] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.927857] env[63418]: value = "task-1245329" [ 954.927857] env[63418]: _type = "Task" [ 954.927857] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.937606] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245329, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.939417] env[63418]: DEBUG oslo_vmware.api [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 954.939417] env[63418]: value = "task-1245330" [ 954.939417] env[63418]: _type = "Task" [ 954.939417] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.951233] env[63418]: DEBUG oslo_vmware.api [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.066819] env[63418]: DEBUG nova.scheduler.client.report [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 955.261071] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.290572] env[63418]: DEBUG nova.network.neutron [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 955.403840] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.443369] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245329, 'name': CreateVM_Task, 'duration_secs': 0.425736} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.447142] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.448038] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.448835] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.449276] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.450052] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85612c9c-f9e1-4532-a63b-98a5116090c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.455934] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 955.455934] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e5b953-e69d-5b57-5820-bc9b9befb408" [ 955.455934] env[63418]: _type = "Task" [ 955.455934] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.459511] env[63418]: DEBUG oslo_vmware.api [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245330, 'name': PowerOffVM_Task, 'duration_secs': 0.249952} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.463169] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.463377] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.463649] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce1f7ab2-9153-4424-8ebd-cfecb50205b6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.472515] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e5b953-e69d-5b57-5820-bc9b9befb408, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.497634] env[63418]: DEBUG nova.network.neutron [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Updating instance_info_cache with network_info: [{"id": "bae1fde8-1485-472f-9c86-34b47e93ee43", "address": "fa:16:3e:d2:4c:4a", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbae1fde8-14", "ovs_interfaceid": "bae1fde8-1485-472f-9c86-34b47e93ee43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.548797] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.549213] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.549520] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Deleting the datastore file [datastore2] bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.550369] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69ec820c-e435-47aa-aa28-4f343df57f3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.558008] env[63418]: DEBUG oslo_vmware.api [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for the task: (returnval){ [ 955.558008] env[63418]: value = "task-1245332" [ 955.558008] env[63418]: _type = "Task" [ 955.558008] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.568435] env[63418]: DEBUG oslo_vmware.api [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.572457] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.573014] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 955.575878] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.882s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.758287] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.904430] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.970754] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e5b953-e69d-5b57-5820-bc9b9befb408, 'name': SearchDatastore_Task, 'duration_secs': 0.091988} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.971389] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.971736] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.972170] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.972609] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.972877] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.973259] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fcce27c6-407c-4739-b2ba-879475413db3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.992533] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.993465] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.994344] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fc9ccfe-b8c3-4993-a47f-b370053be750 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.000825] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-b0887bcd-7ba7-4c0e-8d50-886e27d37649" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.001344] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Instance network_info: |[{"id": "bae1fde8-1485-472f-9c86-34b47e93ee43", "address": "fa:16:3e:d2:4c:4a", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbae1fde8-14", "ovs_interfaceid": "bae1fde8-1485-472f-9c86-34b47e93ee43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 956.001835] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:4c:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bae1fde8-1485-472f-9c86-34b47e93ee43', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.010129] env[63418]: DEBUG oslo.service.loopingcall [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.011972] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.012491] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 956.012491] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cbd25a-cc2b-2068-f2be-c7f5c6cf3041" [ 956.012491] env[63418]: _type = "Task" [ 956.012491] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.012750] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98f036a7-7311-4dd7-9a30-4d9d099f3844 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.045021] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cbd25a-cc2b-2068-f2be-c7f5c6cf3041, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.046149] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.046149] env[63418]: value = "task-1245333" [ 956.046149] env[63418]: _type = "Task" [ 956.046149] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.067716] env[63418]: DEBUG oslo_vmware.api [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Task: {'id': task-1245332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.372759} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.068068] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.068352] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.068590] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.068789] env[63418]: INFO nova.compute.manager [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 956.069054] env[63418]: DEBUG oslo.service.loopingcall [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.069275] env[63418]: DEBUG nova.compute.manager [-] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 956.069376] env[63418]: DEBUG nova.network.neutron [-] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.078875] env[63418]: DEBUG nova.compute.utils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.087745] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 956.088207] env[63418]: DEBUG nova.network.neutron [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 956.135329] env[63418]: DEBUG nova.policy [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '674a82efba9d4d8ca86e52640f9a5589', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e9f537407b84d50a49600de59e72c86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 956.197220] env[63418]: DEBUG nova.compute.manager [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Received event network-changed-bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 956.197220] env[63418]: DEBUG nova.compute.manager [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Refreshing instance network info cache due to event network-changed-bae1fde8-1485-472f-9c86-34b47e93ee43. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 956.197767] env[63418]: DEBUG oslo_concurrency.lockutils [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] Acquiring lock "refresh_cache-b0887bcd-7ba7-4c0e-8d50-886e27d37649" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.198102] env[63418]: DEBUG oslo_concurrency.lockutils [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] Acquired lock "refresh_cache-b0887bcd-7ba7-4c0e-8d50-886e27d37649" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.198399] env[63418]: DEBUG nova.network.neutron [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Refreshing network info cache for port bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 956.263298] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.322773] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfb06f6-6909-48df-8159-077879eba5ce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.331938] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e38b1af-139b-440d-b066-c090bea1ba17 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.365512] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5344b4d0-80f5-4ad5-8330-157ac6557bdd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.375419] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a2fd37-bddf-4307-8dcd-b04ce443f759 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.390526] env[63418]: DEBUG nova.compute.provider_tree [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.403718] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.475642] env[63418]: DEBUG nova.network.neutron [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Successfully created port: 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.551657] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52cbd25a-cc2b-2068-f2be-c7f5c6cf3041, 'name': SearchDatastore_Task, 'duration_secs': 0.092665} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.556046] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91644dd2-8e3c-4452-90d8-588522d6eb79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.564938] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245333, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.566720] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 956.566720] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b272d-b4a5-44a4-6bcb-13162cd4cf6d" [ 956.566720] env[63418]: _type = "Task" [ 956.566720] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.578364] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b272d-b4a5-44a4-6bcb-13162cd4cf6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.589688] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 956.611525] env[63418]: DEBUG nova.compute.manager [req-4452aeb6-ec66-4d79-9d1d-c0591bc0bd39 req-6f63bed6-7670-41ab-b262-0325db82a9a8 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Received event network-vif-deleted-87ac533a-e3c0-46b3-984b-b7631bdc3e22 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 956.611793] env[63418]: INFO nova.compute.manager [req-4452aeb6-ec66-4d79-9d1d-c0591bc0bd39 req-6f63bed6-7670-41ab-b262-0325db82a9a8 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Neutron deleted interface 87ac533a-e3c0-46b3-984b-b7631bdc3e22; detaching it from the instance and deleting it from the info cache [ 956.611993] env[63418]: DEBUG nova.network.neutron [req-4452aeb6-ec66-4d79-9d1d-c0591bc0bd39 req-6f63bed6-7670-41ab-b262-0325db82a9a8 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.761721] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.894184] env[63418]: DEBUG nova.scheduler.client.report [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 956.897838] env[63418]: DEBUG nova.network.neutron [-] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.919495] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245327, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.812133} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.919900] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea/OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea.vmdk to [datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk. [ 956.920043] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Cleaning up location [datastore1] OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 956.920213] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a647c356-2a34-410e-a67d-ad50c697dbea {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.920721] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f09c9fad-24b6-45bb-a43a-cb01acf5ae38 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.929969] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 956.929969] env[63418]: value = "task-1245334" [ 956.929969] env[63418]: _type = "Task" [ 956.929969] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.941939] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245334, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.062507] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245333, 'name': CreateVM_Task, 'duration_secs': 0.578907} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.062698] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.063423] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.063594] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.063919] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.064831] env[63418]: DEBUG nova.network.neutron [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Updated VIF entry in instance network info cache for port bae1fde8-1485-472f-9c86-34b47e93ee43. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.065175] env[63418]: DEBUG nova.network.neutron [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Updating instance_info_cache with network_info: [{"id": "bae1fde8-1485-472f-9c86-34b47e93ee43", "address": "fa:16:3e:d2:4c:4a", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbae1fde8-14", "ovs_interfaceid": "bae1fde8-1485-472f-9c86-34b47e93ee43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.066367] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae1721e3-2f8c-408d-87ce-89fe80194520 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.080616] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527b272d-b4a5-44a4-6bcb-13162cd4cf6d, 'name': SearchDatastore_Task, 'duration_secs': 0.059882} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.080923] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 957.080923] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5243029c-1556-d2fd-80ad-4051a9003d7d" [ 957.080923] env[63418]: _type = "Task" [ 957.080923] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.081801] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.082094] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91/5dac16e3-06a9-443a-90c8-9aacdd23fd91.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.082397] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8e4cca9-ddde-43da-8a25-6b733c506caa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.097522] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5243029c-1556-d2fd-80ad-4051a9003d7d, 'name': SearchDatastore_Task, 'duration_secs': 0.009507} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.099275] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.099517] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.099754] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.099923] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.100128] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.100483] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 957.100483] env[63418]: value = "task-1245335" [ 957.100483] env[63418]: _type = "Task" [ 957.100483] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.100693] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48e71eb3-82e1-4ca6-a74a-74a379511642 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.110731] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245335, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.116428] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ffb05750-8c5d-411f-a380-79adec1b6d94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.121340] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.121530] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.122335] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ed690d0-28f2-479b-9351-0638b7d48932 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.130202] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec463ff-4bf9-48a7-8504-b2b60fc46012 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.142700] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 957.142700] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521a2723-6f8e-8d64-4775-36be3eb6ced2" [ 957.142700] env[63418]: _type = "Task" [ 957.142700] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.152130] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521a2723-6f8e-8d64-4775-36be3eb6ced2, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.152932] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13685f18-d21a-44f2-ac45-ec74a146ccc0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.165142] env[63418]: DEBUG nova.compute.manager [req-4452aeb6-ec66-4d79-9d1d-c0591bc0bd39 req-6f63bed6-7670-41ab-b262-0325db82a9a8 service nova] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Detach interface failed, port_id=87ac533a-e3c0-46b3-984b-b7631bdc3e22, reason: Instance bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 957.169361] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 957.169361] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e33056-7547-1f3b-f6f5-d18ce9018ab1" [ 957.169361] env[63418]: _type = "Task" [ 957.169361] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.176843] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e33056-7547-1f3b-f6f5-d18ce9018ab1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.260640] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.408051] env[63418]: INFO nova.compute.manager [-] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Took 1.34 seconds to deallocate network for instance. [ 957.443300] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245334, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.036756} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.443300] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.443300] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.443442] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk to [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.443978] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-632eacab-3e48-48ef-94e0-049b285219d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.455077] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 957.455077] env[63418]: value = "task-1245336" [ 957.455077] env[63418]: _type = "Task" [ 957.455077] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.466274] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.570141] env[63418]: DEBUG oslo_concurrency.lockutils [req-f105c3c5-e5bd-4da5-b97d-a56a9fc9b813 req-06a23b32-dace-4e3f-b679-5cf310ff9259 service nova] Releasing lock "refresh_cache-b0887bcd-7ba7-4c0e-8d50-886e27d37649" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.600460] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 957.611429] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245335, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495713} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.611695] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91/5dac16e3-06a9-443a-90c8-9aacdd23fd91.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 957.611905] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.612180] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7f07619-a2d5-411a-aab7-28a203c21f23 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.619279] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 957.619279] env[63418]: value = "task-1245337" [ 957.619279] env[63418]: _type = "Task" [ 957.619279] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.628319] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.628621] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.628784] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.628989] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.629230] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.629431] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.629692] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.629892] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.630147] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.630368] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.630569] env[63418]: DEBUG nova.virt.hardware [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.631452] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e792d8e5-342c-4b62-b893-004e5ff87471 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.636933] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.642390] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df76140-718c-484f-9806-3415c1f83bfb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.682667] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e33056-7547-1f3b-f6f5-d18ce9018ab1, 'name': SearchDatastore_Task, 'duration_secs': 0.018636} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.683120] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.683541] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] b0887bcd-7ba7-4c0e-8d50-886e27d37649/b0887bcd-7ba7-4c0e-8d50-886e27d37649.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.683942] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4597ed12-0a3c-4b54-a994-b489056fdab8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.691369] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 957.691369] env[63418]: value = "task-1245338" [ 957.691369] env[63418]: _type = "Task" [ 957.691369] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.701407] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245338, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.763253] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.907625] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.330s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.908085] env[63418]: DEBUG nova.compute.manager [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63418) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5252}} [ 957.911315] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.104s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.911745] env[63418]: DEBUG nova.objects.instance [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'resources' on Instance uuid 26147513-3b25-4cc1-991a-34a724f73711 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.914498] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.969456] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.030588] env[63418]: DEBUG nova.network.neutron [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Successfully updated port: 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 958.130951] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159524} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.132019] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 958.133222] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3313c77-a994-4403-81fe-191cbaf483dc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.154695] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91/5dac16e3-06a9-443a-90c8-9aacdd23fd91.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.155072] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3e6891b-1220-4bc8-a7b9-0e19669f9b2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.174854] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 958.174854] env[63418]: value = "task-1245339" [ 958.174854] env[63418]: _type = "Task" [ 958.174854] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.183598] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245339, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.201289] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245338, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.223219] env[63418]: DEBUG nova.compute.manager [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-vif-plugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 958.223633] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.223706] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.223909] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.224212] env[63418]: DEBUG nova.compute.manager [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] No waiting events found dispatching network-vif-plugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 958.224433] env[63418]: WARNING nova.compute.manager [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received unexpected event network-vif-plugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 for instance with vm_state building and task_state spawning. [ 958.224613] env[63418]: DEBUG nova.compute.manager [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 958.224777] env[63418]: DEBUG nova.compute.manager [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing instance network info cache due to event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 958.224970] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.225125] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.225288] env[63418]: DEBUG nova.network.neutron [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.261714] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.467969] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.474304] env[63418]: INFO nova.scheduler.client.report [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted allocation for migration 35dd30ef-05a7-46f8-bed4-fd19eb2c173a [ 958.534133] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.631257] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06ee386-ba7f-46ed-b203-78772301a250 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.641356] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d39a0e0-54f9-4e65-b5c2-df4ae3507ec5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.680703] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc78f107-1a7f-4d26-940e-437aed96d0c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.690399] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245339, 'name': ReconfigVM_Task, 'duration_secs': 0.485464} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.692869] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91/5dac16e3-06a9-443a-90c8-9aacdd23fd91.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.693620] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b18eb3a7-6d06-41c8-a42c-89d68c44e6d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.699074] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e64d3d-88d0-441f-a1a7-1dfc0be44187 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.713849] env[63418]: DEBUG nova.compute.provider_tree [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.719563] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 958.719563] env[63418]: value = "task-1245340" [ 958.719563] env[63418]: _type = "Task" [ 958.719563] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.719808] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245338, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.735049] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245340, 'name': Rename_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.766890] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.768411] env[63418]: DEBUG nova.network.neutron [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.969201] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.980958] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a73d9e20-3e63-41da-8a21-62b7f7d83233 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.203s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.079920] env[63418]: DEBUG nova.network.neutron [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.205646] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245338, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.224126] env[63418]: DEBUG nova.scheduler.client.report [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 959.252037] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245340, 'name': Rename_Task, 'duration_secs': 0.244289} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.252037] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.252037] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00d9cf41-e915-45c9-9f4a-19acb8e2fcff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.262651] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 959.262651] env[63418]: value = "task-1245341" [ 959.262651] env[63418]: _type = "Task" [ 959.262651] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.278714] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.287872] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245341, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.469451] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.537755] env[63418]: DEBUG nova.objects.instance [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'flavor' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.582699] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ba6fa21-8e21-4a31-a409-0fac78f85667 req-72c393d0-fb9b-4e3d-8ed7-132d249013c7 service nova] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.583100] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.583289] env[63418]: DEBUG nova.network.neutron [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 959.707178] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245338, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.738041] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.740516] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.587s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.762329] env[63418]: INFO nova.scheduler.client.report [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocations for instance 26147513-3b25-4cc1-991a-34a724f73711 [ 959.767569] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.782201] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245341, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.969581] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.044480] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.044761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.045057] env[63418]: DEBUG nova.network.neutron [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.045347] env[63418]: DEBUG nova.objects.instance [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'info_cache' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.122367] env[63418]: DEBUG nova.network.neutron [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 960.205785] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245338, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.110334} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.208446] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] b0887bcd-7ba7-4c0e-8d50-886e27d37649/b0887bcd-7ba7-4c0e-8d50-886e27d37649.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.208727] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.209044] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a3dc163-bd34-4ff6-bab0-e4734a3ff0a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.216974] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 960.216974] env[63418]: value = "task-1245342" [ 960.216974] env[63418]: _type = "Task" [ 960.216974] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.225195] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.263749] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.282251] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.282797] env[63418]: DEBUG oslo_concurrency.lockutils [None req-06e51e32-605b-4f2f-8537-7f2ca261523e tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "26147513-3b25-4cc1-991a-34a724f73711" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.440s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.293360] env[63418]: DEBUG nova.network.neutron [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.467602] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245336, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.708419} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.467887] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/69a22df6-0b8e-4d81-ab99-3f97aaaed86e/69a22df6-0b8e-4d81-ab99-3f97aaaed86e.vmdk to [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.468693] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dddcb7-d7d6-4f04-a50f-bd67abd91faa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.490541] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.490886] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85723afb-8cb4-4c55-8e02-9c9e7b986156 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.510209] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 960.510209] env[63418]: value = "task-1245343" [ 960.510209] env[63418]: _type = "Task" [ 960.510209] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.518988] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.549987] env[63418]: DEBUG nova.objects.base [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Object Instance<7b0c70aa-e2bc-4131-97b4-4e53a378940a> lazy-loaded attributes: flavor,info_cache {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 960.726178] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229116} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.726460] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.727239] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c653e74e-e86e-47cb-871b-31cbb9abd725 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.748766] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] b0887bcd-7ba7-4c0e-8d50-886e27d37649/b0887bcd-7ba7-4c0e-8d50-886e27d37649.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.749087] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ac70af0-b836-44a1-81e6-078e56a5b721 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.777890] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 960.777890] env[63418]: value = "task-1245344" [ 960.777890] env[63418]: _type = "Task" [ 960.777890] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.778151] env[63418]: DEBUG oslo_vmware.api [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245328, 'name': ReconfigVM_Task, 'duration_secs': 5.902757} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.778755] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.778982] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Reconfigured VM to detach interface {{(pid=63418) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 960.786125] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.786982] env[63418]: WARNING nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 960.787135] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance bf091892-fefb-49dd-9416-708a06b35798 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.787258] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance dcb52a03-8f82-4b33-8dc6-3924140db8d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.787374] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.787486] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance cbcbaf4f-ee24-4072-83ae-ffde59478928 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.787599] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 0f99b32a-0125-4df0-919c-e5456b4ae4a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.787735] env[63418]: WARNING nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 23af1d2e-f8ec-4c09-a070-9719b593f83a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 960.787851] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7b0c70aa-e2bc-4131-97b4-4e53a378940a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.787961] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.788080] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 5dac16e3-06a9-443a-90c8-9aacdd23fd91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.788191] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance b0887bcd-7ba7-4c0e-8d50-886e27d37649 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.788299] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance d52db46b-2461-4bd0-be57-d414250aac7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.788487] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 960.788622] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 960.795975] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.796278] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance network_info: |[{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 960.796550] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245344, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.796934] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:76:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cc10d6a-03cf-43e3-84f7-9160f8626ad4', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 960.804623] env[63418]: DEBUG oslo.service.loopingcall [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.807467] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 960.808070] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00a3d277-11a5-474c-9492-5e0eba43abaf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.835224] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 960.835224] env[63418]: value = "task-1245345" [ 960.835224] env[63418]: _type = "Task" [ 960.835224] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.846025] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245345, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.967781] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c51b613-cda0-469e-a69a-07e2c94e8409 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.976300] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7443e5-c15c-4fd6-97ce-4fff07095afb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.015071] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893ccc9e-ac99-43f3-accd-3270dd8e8654 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.028046] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86e52e4-d1cf-4660-a3d9-301de4cf2ed9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.032899] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245343, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.043699] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.288913] env[63418]: DEBUG oslo_vmware.api [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245341, 'name': PowerOnVM_Task, 'duration_secs': 1.766313} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.291486] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.291685] env[63418]: INFO nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Took 9.75 seconds to spawn the instance on the hypervisor. [ 961.292316] env[63418]: DEBUG nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 961.292971] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d6d9e4-9710-41a7-87e0-1efa885b3524 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.298560] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245344, 'name': ReconfigVM_Task, 'duration_secs': 0.309667} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.299173] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Reconfigured VM instance instance-0000005d to attach disk [datastore1] b0887bcd-7ba7-4c0e-8d50-886e27d37649/b0887bcd-7ba7-4c0e-8d50-886e27d37649.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.299792] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6c0f027-5f07-452b-b352-61575aa3d94f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.311018] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 961.311018] env[63418]: value = "task-1245346" [ 961.311018] env[63418]: _type = "Task" [ 961.311018] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.316469] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245346, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.346024] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245345, 'name': CreateVM_Task, 'duration_secs': 0.377472} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.346024] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 961.350528] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.350730] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.351126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.352136] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12aeecbc-44e1-42fe-aef9-ef65290a1a95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.362314] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 961.362314] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521c3792-a706-5c2c-dc09-dc1c512c8c31" [ 961.362314] env[63418]: _type = "Task" [ 961.362314] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.365325] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.365561] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.375857] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521c3792-a706-5c2c-dc09-dc1c512c8c31, 'name': SearchDatastore_Task, 'duration_secs': 0.011442} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.376699] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.376991] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.377287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.377450] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.377639] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.378131] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2fc23d0-47b5-4788-9347-48263c462430 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.386426] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.390018] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.390018] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e170b379-db17-4285-9d72-63ff6b1b1fda {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.392625] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 961.392625] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5223dccb-6ec1-b4a8-d2a4-83c761b810b3" [ 961.392625] env[63418]: _type = "Task" [ 961.392625] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.400926] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5223dccb-6ec1-b4a8-d2a4-83c761b810b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.509766] env[63418]: DEBUG nova.network.neutron [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [{"id": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "address": "fa:16:3e:e6:0e:e1", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0fa119b-63", "ovs_interfaceid": "d0fa119b-63f3-4b54-8592-195a043ff0ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.521096] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245343, 'name': ReconfigVM_Task, 'duration_secs': 0.531523} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.522052] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8/6465fb5c-7bc9-4197-b3fb-bad2c000a5f8.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.522754] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd657a36-8950-47a1-9d4c-0dad8a639419 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.529432] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 961.529432] env[63418]: value = "task-1245347" [ 961.529432] env[63418]: _type = "Task" [ 961.529432] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.538877] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245347, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.546229] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 961.817323] env[63418]: INFO nova.compute.manager [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Took 20.35 seconds to build instance. [ 961.822319] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245346, 'name': Rename_Task, 'duration_secs': 0.137061} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.822433] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.823034] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c16ee83b-1f56-4119-b1b1-5169158d1e74 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.829159] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 961.829159] env[63418]: value = "task-1245348" [ 961.829159] env[63418]: _type = "Task" [ 961.829159] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.837813] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245348, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.870669] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 961.904604] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5223dccb-6ec1-b4a8-d2a4-83c761b810b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008047} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.904783] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f28facf8-8d7b-4324-87e6-0aa24915f384 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.910905] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 961.910905] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520be1b8-c1db-08d5-492e-5a7e7596f527" [ 961.910905] env[63418]: _type = "Task" [ 961.910905] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.922213] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520be1b8-c1db-08d5-492e-5a7e7596f527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.013175] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.040814] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245347, 'name': Rename_Task, 'duration_secs': 0.391895} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.041056] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.041334] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5c55777-95b3-4fd2-8137-fbb0fe7582fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.049814] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 962.049814] env[63418]: value = "task-1245349" [ 962.049814] env[63418]: _type = "Task" [ 962.049814] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.053205] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 962.053364] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.313s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.053606] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.096s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.053789] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.055648] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.141s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.055857] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.062308] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245349, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.072996] env[63418]: INFO nova.scheduler.client.report [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Deleted allocations for instance bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5 [ 962.074957] env[63418]: INFO nova.scheduler.client.report [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted allocations for instance 23af1d2e-f8ec-4c09-a070-9719b593f83a [ 962.179727] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.180028] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.180288] env[63418]: DEBUG nova.network.neutron [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.319643] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9da51548-149a-4959-a2bf-a73465f9f1e1 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.864s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.340794] env[63418]: DEBUG oslo_vmware.api [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245348, 'name': PowerOnVM_Task, 'duration_secs': 0.445798} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.340794] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.341658] env[63418]: INFO nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Took 8.44 seconds to spawn the instance on the hypervisor. [ 962.341888] env[63418]: DEBUG nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 962.342867] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19b661a-9140-49e6-9ecd-bccce9e01bfa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.397640] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.397640] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.399073] env[63418]: INFO nova.compute.claims [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.425413] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520be1b8-c1db-08d5-492e-5a7e7596f527, 'name': SearchDatastore_Task, 'duration_secs': 0.011249} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.425413] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.425413] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.425413] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ad06787-76a6-4ce5-91fa-c432553f0cfd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.431347] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 962.431347] env[63418]: value = "task-1245350" [ 962.431347] env[63418]: _type = "Task" [ 962.431347] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.441787] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.564857] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245349, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.586885] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a95f7af2-77d5-4abc-88d1-56aebadb418d tempest-ServerRescueTestJSON-268293277 tempest-ServerRescueTestJSON-268293277-project-member] Lock "bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.712s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.588774] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1fa57ae6-6aab-4f2d-9494-a58ea85ea1d7 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "23af1d2e-f8ec-4c09-a070-9719b593f83a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.050s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.865339] env[63418]: INFO nova.compute.manager [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Took 20.75 seconds to build instance. [ 962.944334] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245350, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.974508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.974508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.974508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.974508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.974508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.976198] env[63418]: INFO nova.compute.manager [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Terminating instance [ 963.021890] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.022557] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dfa803b-8e32-44df-87d0-f94c941f770f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.030271] env[63418]: DEBUG oslo_vmware.api [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 963.030271] env[63418]: value = "task-1245351" [ 963.030271] env[63418]: _type = "Task" [ 963.030271] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.044588] env[63418]: DEBUG oslo_vmware.api [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.045194] env[63418]: DEBUG nova.compute.manager [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Stashing vm_state: active {{(pid=63418) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 963.062566] env[63418]: DEBUG oslo_vmware.api [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245349, 'name': PowerOnVM_Task, 'duration_secs': 0.671257} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.064018] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.186804] env[63418]: DEBUG nova.compute.manager [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 963.188111] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fa2b35-6179-4c01-8dd9-cdf695eecb39 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.309545] env[63418]: INFO nova.network.neutron [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Port f1f171f0-8f67-4b5b-a9c8-91014404b4e2 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 963.309839] env[63418]: DEBUG nova.network.neutron [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.369864] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b323e3c3-08bb-4dd5-822c-9724f6fdf9fb tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.259s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.447240] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539994} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.450918] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.451169] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.451629] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4a61f48-f882-42cc-a121-6ee4dc5cd387 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.459424] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 963.459424] env[63418]: value = "task-1245352" [ 963.459424] env[63418]: _type = "Task" [ 963.459424] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.470886] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245352, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.479997] env[63418]: DEBUG nova.compute.manager [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 963.480258] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.481116] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225a03f5-70e2-45fe-ac6a-cd3bb73a8054 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.491954] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.492243] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d49f7f9c-03fd-4bbb-82bf-58035dc0696c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.498976] env[63418]: DEBUG oslo_vmware.api [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 963.498976] env[63418]: value = "task-1245353" [ 963.498976] env[63418]: _type = "Task" [ 963.498976] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.511786] env[63418]: DEBUG oslo_vmware.api [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.542568] env[63418]: DEBUG oslo_vmware.api [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245351, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.577907] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.633156] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6bbb49-de89-432c-934e-14c365fc3fe1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.642340] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed8e04b-84ee-405a-b9e1-62277e3bcd48 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.677645] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd61337-5f7f-42db-b3cd-949ae1016a31 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.686505] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06537c15-af83-4a5a-88c9-46b343977adc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.705496] env[63418]: DEBUG nova.compute.provider_tree [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.710422] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d2d9aa7-f448-472c-91c8-3c5c50388731 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.822s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.817103] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.983115] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0c016d4e-ca34-4831-a567-e794012681db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.983115] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0c016d4e-ca34-4831-a567-e794012681db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.984276] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245352, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116646} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.984777] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.985873] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3189bbc9-4407-48b4-ab16-4202b48b5bcd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.009545] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.013822] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed3cc6ac-b0b8-4699-bfa8-18598e788114 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.031269] env[63418]: DEBUG nova.compute.manager [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 964.032131] env[63418]: DEBUG nova.compute.manager [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing instance network info cache due to event network-changed-9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 964.032239] env[63418]: DEBUG oslo_concurrency.lockutils [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] Acquiring lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.032500] env[63418]: DEBUG oslo_concurrency.lockutils [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] Acquired lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.032767] env[63418]: DEBUG nova.network.neutron [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Refreshing network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.049293] env[63418]: DEBUG oslo_vmware.api [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245353, 'name': PowerOffVM_Task, 'duration_secs': 0.276404} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.049293] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 964.049293] env[63418]: value = "task-1245354" [ 964.049293] env[63418]: _type = "Task" [ 964.049293] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.049293] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.049293] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.049293] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7ec640c-00fc-4edc-b67d-987004e38a8c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.056603] env[63418]: DEBUG oslo_vmware.api [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245351, 'name': PowerOnVM_Task, 'duration_secs': 0.710699} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.056603] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.056603] env[63418]: DEBUG nova.compute.manager [None req-b5fc972b-b8ab-47b4-bbe0-4924b36044af tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 964.057466] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c421a4-a875-4bcf-831f-930dd3d1cdf6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.063802] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245354, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.123937] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.124366] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.124452] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleting the datastore file [datastore2] dcb52a03-8f82-4b33-8dc6-3924140db8d3 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.124722] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-069e42f6-be87-49f3-b2d5-f02e8b6d5842 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.131021] env[63418]: DEBUG oslo_vmware.api [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 964.131021] env[63418]: value = "task-1245356" [ 964.131021] env[63418]: _type = "Task" [ 964.131021] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.139128] env[63418]: DEBUG oslo_vmware.api [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245356, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.211075] env[63418]: DEBUG nova.scheduler.client.report [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 964.324980] env[63418]: DEBUG oslo_concurrency.lockutils [None req-420b201c-02eb-4d79-bc63-1a3787c7a9d7 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.154s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.485726] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 964.557859] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245354, 'name': ReconfigVM_Task, 'duration_secs': 0.291496} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.558174] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Reconfigured VM instance instance-0000005e to attach disk [datastore2] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.558817] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1fcbd22-5cb5-43e0-890a-09faa089b8b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.567020] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 964.567020] env[63418]: value = "task-1245357" [ 964.567020] env[63418]: _type = "Task" [ 964.567020] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.576389] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245357, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.645091] env[63418]: DEBUG oslo_vmware.api [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245356, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161302} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.645609] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.646213] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 964.646312] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 964.646549] env[63418]: INFO nova.compute.manager [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 964.646931] env[63418]: DEBUG oslo.service.loopingcall [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.647346] env[63418]: DEBUG nova.compute.manager [-] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 964.647495] env[63418]: DEBUG nova.network.neutron [-] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.677027] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-0f99b32a-0125-4df0-919c-e5456b4ae4a5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.677390] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-0f99b32a-0125-4df0-919c-e5456b4ae4a5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.677804] env[63418]: DEBUG nova.objects.instance [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'flavor' on Instance uuid 0f99b32a-0125-4df0-919c-e5456b4ae4a5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.720651] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.721190] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 964.727629] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.150s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.896805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.896805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.896805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.896805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.896805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.906795] env[63418]: INFO nova.compute.manager [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Terminating instance [ 965.011216] env[63418]: DEBUG nova.network.neutron [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updated VIF entry in instance network info cache for port 9b1f9284-9bb8-49b0-80f1-c2154e6ba534. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.011624] env[63418]: DEBUG nova.network.neutron [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [{"id": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "address": "fa:16:3e:7f:f0:1d", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b1f9284-9b", "ovs_interfaceid": "9b1f9284-9bb8-49b0-80f1-c2154e6ba534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.013797] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.076103] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245357, 'name': Rename_Task, 'duration_secs': 0.145782} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.076710] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.077099] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de665e76-f340-491c-b200-331d86bb52a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.085950] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 965.085950] env[63418]: value = "task-1245358" [ 965.085950] env[63418]: _type = "Task" [ 965.085950] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.097720] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245358, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.236853] env[63418]: INFO nova.compute.claims [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 965.242327] env[63418]: DEBUG nova.compute.utils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.243628] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 965.243798] env[63418]: DEBUG nova.network.neutron [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 965.293530] env[63418]: DEBUG nova.policy [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ff64612530b451fb41100a5aa601be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d52f1fceb24234a8d967038b43c857', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.316704] env[63418]: DEBUG nova.compute.manager [req-2cf4081c-16ae-42dd-ae7f-2a25ba224c6d req-8ac9e926-fd97-4692-abdf-ade5b56b4aef service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Received event network-vif-deleted-65da165e-6d27-4c6a-9c82-0186992d1648 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 965.316928] env[63418]: INFO nova.compute.manager [req-2cf4081c-16ae-42dd-ae7f-2a25ba224c6d req-8ac9e926-fd97-4692-abdf-ade5b56b4aef service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Neutron deleted interface 65da165e-6d27-4c6a-9c82-0186992d1648; detaching it from the instance and deleting it from the info cache [ 965.317142] env[63418]: DEBUG nova.network.neutron [req-2cf4081c-16ae-42dd-ae7f-2a25ba224c6d req-8ac9e926-fd97-4692-abdf-ade5b56b4aef service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.415225] env[63418]: DEBUG nova.compute.manager [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 965.415941] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.417588] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c59c5a-508a-4bb7-af3c-f83768dbbb20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.426883] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.427221] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ea9aad3-dc0d-425e-a94f-7811947ad323 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.435148] env[63418]: DEBUG oslo_vmware.api [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 965.435148] env[63418]: value = "task-1245359" [ 965.435148] env[63418]: _type = "Task" [ 965.435148] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.443835] env[63418]: DEBUG nova.network.neutron [-] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.452825] env[63418]: DEBUG oslo_vmware.api [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.505307] env[63418]: DEBUG nova.objects.instance [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'pci_requests' on Instance uuid 0f99b32a-0125-4df0-919c-e5456b4ae4a5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.517033] env[63418]: DEBUG oslo_concurrency.lockutils [req-de3007e9-ae6d-4f52-9217-5063921b31d9 req-d665989b-8b84-428d-b71a-f69c1ccd2da9 service nova] Releasing lock "refresh_cache-039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.600877] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245358, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.610158] env[63418]: DEBUG nova.network.neutron [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Successfully created port: 492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.752413] env[63418]: INFO nova.compute.resource_tracker [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating resource usage from migration 9856ef27-c1c8-4c00-8c0e-e947b31de4f1 [ 965.756429] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 965.823192] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36a16d2a-1b36-448f-a9c7-5778f19cf1da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.838234] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ac8d18-8d2e-49ee-89ed-14ffade6d954 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.875305] env[63418]: DEBUG nova.compute.manager [req-2cf4081c-16ae-42dd-ae7f-2a25ba224c6d req-8ac9e926-fd97-4692-abdf-ade5b56b4aef service nova] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Detach interface failed, port_id=65da165e-6d27-4c6a-9c82-0186992d1648, reason: Instance dcb52a03-8f82-4b33-8dc6-3924140db8d3 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 965.948337] env[63418]: INFO nova.compute.manager [-] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Took 1.30 seconds to deallocate network for instance. [ 965.949169] env[63418]: DEBUG oslo_vmware.api [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245359, 'name': PowerOffVM_Task, 'duration_secs': 0.198451} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.953523] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.953702] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.954785] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9264c526-b436-41cd-82a4-810752c508b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.996927] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc49f04e-82d1-4cbb-9a4a-d7e1955e481f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.006015] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90262c7-2072-4e40-abc3-8893e76ae363 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.009871] env[63418]: DEBUG nova.objects.base [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Object Instance<0f99b32a-0125-4df0-919c-e5456b4ae4a5> lazy-loaded attributes: flavor,pci_requests {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 966.010105] env[63418]: DEBUG nova.network.neutron [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.043930] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b846275-1948-4ee2-9e38-21ce4b985f1e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.046977] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.047243] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.047431] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore1] b0887bcd-7ba7-4c0e-8d50-886e27d37649 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.047742] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d75548c-26bc-4cb6-8230-8b393d2fea0c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.057777] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63193d1c-df42-4782-8d9b-f51f0fc8ed29 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.061799] env[63418]: DEBUG oslo_vmware.api [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 966.061799] env[63418]: value = "task-1245361" [ 966.061799] env[63418]: _type = "Task" [ 966.061799] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.074035] env[63418]: DEBUG nova.compute.provider_tree [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.080028] env[63418]: DEBUG oslo_vmware.api [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245361, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.095066] env[63418]: DEBUG oslo_vmware.api [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245358, 'name': PowerOnVM_Task, 'duration_secs': 0.520044} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.095362] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.095604] env[63418]: INFO nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Took 8.49 seconds to spawn the instance on the hypervisor. [ 966.095804] env[63418]: DEBUG nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 966.096655] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dc3223-7052-4382-b417-0a2036990f09 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.163085] env[63418]: DEBUG nova.policy [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2cca4ff5894585afa66ab960f5370b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4297b53faeab40dfa5de863ad4030800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.433199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.433199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.433199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.433199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.433199] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.436917] env[63418]: INFO nova.compute.manager [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Terminating instance [ 966.460979] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.573808] env[63418]: DEBUG oslo_vmware.api [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245361, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265098} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.574261] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.574527] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.574776] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.574917] env[63418]: INFO nova.compute.manager [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Took 1.16 seconds to destroy the instance on the hypervisor. [ 966.575193] env[63418]: DEBUG oslo.service.loopingcall [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.575412] env[63418]: DEBUG nova.compute.manager [-] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 966.575650] env[63418]: DEBUG nova.network.neutron [-] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 966.578163] env[63418]: DEBUG nova.scheduler.client.report [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 966.619181] env[63418]: INFO nova.compute.manager [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Took 20.70 seconds to build instance. [ 966.764275] env[63418]: DEBUG nova.compute.manager [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-changed-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 966.764570] env[63418]: DEBUG nova.compute.manager [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing instance network info cache due to event network-changed-34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 966.764761] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.764876] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.765886] env[63418]: DEBUG nova.network.neutron [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.768780] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 966.808668] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.808933] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.809096] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.809924] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.809924] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.809924] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.809924] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.809924] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.810399] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.810668] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.810906] env[63418]: DEBUG nova.virt.hardware [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.812318] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df54dee-089f-4817-a8b4-ccafa171f9e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.822825] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3427ddaf-ccae-4c5c-adff-69bfd93eaf87 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.948241] env[63418]: DEBUG nova.compute.manager [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 966.948241] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.951706] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715cba17-4a4d-467a-890d-5d6d64e93685 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.958645] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.958908] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-086d9feb-9a39-43c3-b154-cd0ff01ca729 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.965454] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 966.965454] env[63418]: value = "task-1245362" [ 966.965454] env[63418]: _type = "Task" [ 966.965454] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.976684] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.082883] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.355s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.083123] env[63418]: INFO nova.compute.manager [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Migrating [ 967.091150] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.077s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.094077] env[63418]: INFO nova.compute.claims [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.125209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-894fadf3-d73b-418e-b22e-bf3e5332d672 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.214s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.215912] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 967.479951] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245362, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.524515] env[63418]: DEBUG nova.network.neutron [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Successfully updated port: 492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.529703] env[63418]: DEBUG nova.network.neutron [-] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.613508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.613724] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.613903] env[63418]: DEBUG nova.network.neutron [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.620600] env[63418]: DEBUG nova.compute.manager [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Received event network-vif-deleted-bae1fde8-1485-472f-9c86-34b47e93ee43 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 967.622017] env[63418]: DEBUG nova.compute.manager [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Received event network-vif-plugged-492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 967.622017] env[63418]: DEBUG oslo_concurrency.lockutils [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] Acquiring lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.622017] env[63418]: DEBUG oslo_concurrency.lockutils [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.622017] env[63418]: DEBUG oslo_concurrency.lockutils [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.622017] env[63418]: DEBUG nova.compute.manager [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] No waiting events found dispatching network-vif-plugged-492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 967.622017] env[63418]: WARNING nova.compute.manager [req-07da3566-a999-4aba-add5-da75c8f0e1b4 req-c326a03a-6272-4570-9845-afbef158e161 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Received unexpected event network-vif-plugged-492342f5-9181-4e37-9a04-ea9419e43523 for instance with vm_state building and task_state spawning. [ 967.722670] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 967.722805] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 967.752279] env[63418]: DEBUG nova.network.neutron [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updated VIF entry in instance network info cache for port 34d14910-009f-4ee8-b718-a43961c430cb. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.752769] env[63418]: DEBUG nova.network.neutron [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.977536] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245362, 'name': PowerOffVM_Task, 'duration_secs': 0.528708} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.979730] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 967.979730] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 967.979730] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1403c61-4210-455f-b52b-ccca97f0622c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.028394] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.028561] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.028714] env[63418]: DEBUG nova.network.neutron [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.033315] env[63418]: INFO nova.compute.manager [-] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Took 1.46 seconds to deallocate network for instance. [ 968.042527] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.042801] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.042993] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleting the datastore file [datastore1] 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.043529] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c06fd94-ba4a-4b5c-8dcd-d75975e913f0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.053798] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 968.053798] env[63418]: value = "task-1245364" [ 968.053798] env[63418]: _type = "Task" [ 968.053798] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.061044] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.190667] env[63418]: DEBUG nova.network.neutron [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Successfully updated port: f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.257805] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d40dc96-6956-4bde-b731-cf8a3b607a17 req-1cfd3597-3eb5-4caf-ba68-144029f6b820 service nova] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.351217] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4a5b42-d7ce-4e98-b808-7ef12d117639 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.361130] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439ee575-5342-4d52-93c9-8958857caeb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.397024] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54811eb5-f96f-434e-a7eb-18a4c2a650a3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.405693] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95e793c-5153-4cfe-8127-c885a33b830f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.423326] env[63418]: DEBUG nova.compute.provider_tree [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.437523] env[63418]: DEBUG nova.network.neutron [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [{"id": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "address": "fa:16:3e:43:d6:a0", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ab35b58-10", "ovs_interfaceid": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.467794] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0bfa5cf-4641-4e6f-b6b8-3ab2e7db13a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.475158] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Suspending the VM {{(pid=63418) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 968.475466] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5df6950d-71e4-401d-9393-f79ce180b2cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.483635] env[63418]: DEBUG oslo_vmware.api [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 968.483635] env[63418]: value = "task-1245365" [ 968.483635] env[63418]: _type = "Task" [ 968.483635] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.496326] env[63418]: DEBUG oslo_vmware.api [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245365, 'name': SuspendVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.540172] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.563616] env[63418]: DEBUG oslo_vmware.api [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19625} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.563965] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.564207] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.564472] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.564709] env[63418]: INFO nova.compute.manager [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 1.62 seconds to destroy the instance on the hypervisor. [ 968.565084] env[63418]: DEBUG oslo.service.loopingcall [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.565286] env[63418]: DEBUG nova.compute.manager [-] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 968.565395] env[63418]: DEBUG nova.network.neutron [-] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 968.576192] env[63418]: DEBUG nova.network.neutron [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.696388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.696736] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.696839] env[63418]: DEBUG nova.network.neutron [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.800218] env[63418]: DEBUG nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-vif-plugged-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 968.800842] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.801487] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.802086] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.802426] env[63418]: DEBUG nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] No waiting events found dispatching network-vif-plugged-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 968.802906] env[63418]: WARNING nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received unexpected event network-vif-plugged-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 for instance with vm_state active and task_state None. [ 968.803368] env[63418]: DEBUG nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 968.803848] env[63418]: DEBUG nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing instance network info cache due to event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 968.804663] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.804663] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.806413] env[63418]: DEBUG nova.network.neutron [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.841417] env[63418]: DEBUG nova.network.neutron [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [{"id": "492342f5-9181-4e37-9a04-ea9419e43523", "address": "fa:16:3e:6e:07:1a", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap492342f5-91", "ovs_interfaceid": "492342f5-9181-4e37-9a04-ea9419e43523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.927054] env[63418]: DEBUG nova.scheduler.client.report [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 968.940371] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.998353] env[63418]: DEBUG oslo_vmware.api [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245365, 'name': SuspendVM_Task} progress is 62%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.246095] env[63418]: WARNING nova.network.neutron [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] f11c9d94-8e8f-4c7d-b518-f8958080b8fa already exists in list: networks containing: ['f11c9d94-8e8f-4c7d-b518-f8958080b8fa']. ignoring it [ 969.344347] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.347020] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Instance network_info: |[{"id": "492342f5-9181-4e37-9a04-ea9419e43523", "address": "fa:16:3e:6e:07:1a", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap492342f5-91", "ovs_interfaceid": "492342f5-9181-4e37-9a04-ea9419e43523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 969.347020] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:07:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '492342f5-9181-4e37-9a04-ea9419e43523', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.354707] env[63418]: DEBUG oslo.service.loopingcall [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.358488] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.358943] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f5d7163-d62f-4f29-928a-c44357e7793b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.381676] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.381676] env[63418]: value = "task-1245366" [ 969.381676] env[63418]: _type = "Task" [ 969.381676] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.397224] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245366, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.435942] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.440306] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 969.444334] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.983s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.444819] env[63418]: DEBUG nova.objects.instance [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lazy-loading 'resources' on Instance uuid dcb52a03-8f82-4b33-8dc6-3924140db8d3 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.499641] env[63418]: DEBUG oslo_vmware.api [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245365, 'name': SuspendVM_Task, 'duration_secs': 0.971704} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.499944] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Suspended the VM {{(pid=63418) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 969.500371] env[63418]: DEBUG nova.compute.manager [None req-8aa26be2-30e9-4372-b6c8-4212f5962c57 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 969.501245] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a78249-3246-4cd3-9b76-1cd9a1a58b07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.611516] env[63418]: DEBUG nova.network.neutron [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "address": "fa:16:3e:be:8e:cc", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1f171f0-8f", "ovs_interfaceid": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.728548] env[63418]: DEBUG nova.network.neutron [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updated VIF entry in instance network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 969.728951] env[63418]: DEBUG nova.network.neutron [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.803412] env[63418]: DEBUG nova.network.neutron [-] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.893207] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245366, 'name': CreateVM_Task, 'duration_secs': 0.346571} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.893207] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 969.893207] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.893207] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.893207] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.893207] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db52b52d-3919-4f03-9ede-9ec755168a9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.899249] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 969.899249] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5223dd38-0b1d-33ea-eaa3-f86d3b34e47b" [ 969.899249] env[63418]: _type = "Task" [ 969.899249] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.909280] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5223dd38-0b1d-33ea-eaa3-f86d3b34e47b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.959906] env[63418]: DEBUG nova.compute.utils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 969.964055] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 969.964055] env[63418]: DEBUG nova.network.neutron [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.037339] env[63418]: DEBUG nova.policy [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea507bad11c3406d880ba47d08a047c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd445600834dd4c7e8022349ee993f3ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.063249] env[63418]: DEBUG nova.compute.manager [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Received event network-changed-492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 970.063459] env[63418]: DEBUG nova.compute.manager [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Refreshing instance network info cache due to event network-changed-492342f5-9181-4e37-9a04-ea9419e43523. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 970.063677] env[63418]: DEBUG oslo_concurrency.lockutils [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] Acquiring lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.063824] env[63418]: DEBUG oslo_concurrency.lockutils [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] Acquired lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.063989] env[63418]: DEBUG nova.network.neutron [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Refreshing network info cache for port 492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.114725] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.115463] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.115647] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.116500] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958f9518-5342-47b4-b779-8845407a5a7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.136282] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.136650] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.136888] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.137044] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.137276] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.137547] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.137890] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.138117] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.138421] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.138603] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.138798] env[63418]: DEBUG nova.virt.hardware [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.147026] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Reconfiguring VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 970.150621] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7c43772-1adb-4a3f-8ab8-cd658fa227ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.167534] env[63418]: DEBUG oslo_vmware.api [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 970.167534] env[63418]: value = "task-1245367" [ 970.167534] env[63418]: _type = "Task" [ 970.167534] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.176305] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1207a69-b3ca-4958-8cad-45c3d64d89a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.179867] env[63418]: DEBUG oslo_vmware.api [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245367, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.184339] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a24f4e-790e-4502-8037-b80056d49e72 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.216205] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9ccc9f-f6f6-42be-8b0f-24abeea40cb1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.223883] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56896b93-1267-47d0-a06d-45ba407fa090 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.237823] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.238201] env[63418]: DEBUG nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-changed-f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 970.238445] env[63418]: DEBUG nova.compute.manager [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing instance network info cache due to event network-changed-f1f171f0-8f67-4b5b-a9c8-91014404b4e2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 970.238721] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.238925] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.239168] env[63418]: DEBUG nova.network.neutron [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Refreshing network info cache for port f1f171f0-8f67-4b5b-a9c8-91014404b4e2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.240561] env[63418]: DEBUG nova.compute.provider_tree [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.308261] env[63418]: INFO nova.compute.manager [-] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Took 1.74 seconds to deallocate network for instance. [ 970.355428] env[63418]: DEBUG nova.network.neutron [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Successfully created port: 050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.410516] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5223dd38-0b1d-33ea-eaa3-f86d3b34e47b, 'name': SearchDatastore_Task, 'duration_secs': 0.042991} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.410848] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.411151] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.411468] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.411648] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.411844] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.412408] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5cc91c1-15d0-4453-a703-cd87e3d63582 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.424436] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.424660] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.425789] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03a3f4ee-a456-4ca2-9c4b-1438dfd4d5ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.433182] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 970.433182] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5260abd3-3a38-8ddd-8278-950b4269ce69" [ 970.433182] env[63418]: _type = "Task" [ 970.433182] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.442967] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.443256] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.448161] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5260abd3-3a38-8ddd-8278-950b4269ce69, 'name': SearchDatastore_Task, 'duration_secs': 0.010919} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.449138] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef183ed6-1ca3-4251-a1c0-627bdc3aecb0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.454394] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 970.454394] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528c7b02-af6c-43dc-c508-a94fa7a97c0a" [ 970.454394] env[63418]: _type = "Task" [ 970.454394] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.463276] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528c7b02-af6c-43dc-c508-a94fa7a97c0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.463762] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 970.471516] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28095e24-2617-42fd-9021-96e28e803440 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.489836] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 0 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.678058] env[63418]: DEBUG oslo_vmware.api [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245367, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.747871] env[63418]: DEBUG nova.scheduler.client.report [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 970.754021] env[63418]: INFO nova.compute.manager [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Resuming [ 970.754021] env[63418]: DEBUG nova.objects.instance [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lazy-loading 'flavor' on Instance uuid 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.815076] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.896745] env[63418]: DEBUG nova.compute.manager [req-056fe1b0-ca21-4699-a5c9-19c7e3c9f65a req-809143f2-d4e6-47f3-a205-0ee26214b2ec service nova] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Received event network-vif-deleted-d0fa119b-63f3-4b54-8592-195a043ff0ee {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 970.949691] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 970.967450] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528c7b02-af6c-43dc-c508-a94fa7a97c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.967450] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.967450] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6/680c94e8-c6d9-4aab-92d4-047a0e70eec6.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.967450] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b135fcb8-cfdc-433b-8b34-40aa9af34624 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.980660] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 970.980660] env[63418]: value = "task-1245368" [ 970.980660] env[63418]: _type = "Task" [ 970.980660] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.991074] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.995973] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.996548] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45d77dde-4a3d-491e-9c78-cc469ae558ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.002018] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 971.002018] env[63418]: value = "task-1245369" [ 971.002018] env[63418]: _type = "Task" [ 971.002018] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.015027] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.103241] env[63418]: DEBUG nova.network.neutron [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updated VIF entry in instance network info cache for port 492342f5-9181-4e37-9a04-ea9419e43523. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.103680] env[63418]: DEBUG nova.network.neutron [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [{"id": "492342f5-9181-4e37-9a04-ea9419e43523", "address": "fa:16:3e:6e:07:1a", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap492342f5-91", "ovs_interfaceid": "492342f5-9181-4e37-9a04-ea9419e43523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.173890] env[63418]: DEBUG nova.network.neutron [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updated VIF entry in instance network info cache for port f1f171f0-8f67-4b5b-a9c8-91014404b4e2. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.174375] env[63418]: DEBUG nova.network.neutron [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "address": "fa:16:3e:be:8e:cc", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1f171f0-8f", "ovs_interfaceid": "f1f171f0-8f67-4b5b-a9c8-91014404b4e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.182093] env[63418]: DEBUG oslo_vmware.api [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245367, 'name': ReconfigVM_Task, 'duration_secs': 0.843317} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.182956] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.183227] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Reconfigured VM to attach interface {{(pid=63418) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 971.256512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.262508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.722s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.262508] env[63418]: DEBUG nova.objects.instance [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid b0887bcd-7ba7-4c0e-8d50-886e27d37649 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.283930] env[63418]: INFO nova.scheduler.client.report [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted allocations for instance dcb52a03-8f82-4b33-8dc6-3924140db8d3 [ 971.474908] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.478488] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 971.490871] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245368, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.503057] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.503312] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.503474] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.503659] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.503811] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.503957] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.504186] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.504351] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.504623] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.504786] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.505026] env[63418]: DEBUG nova.virt.hardware [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.505899] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ff8e6e-1e8d-4d3f-b3f1-fdd539f52d41 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.516571] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245369, 'name': PowerOffVM_Task, 'duration_secs': 0.189164} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.518601] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.518803] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 17 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 971.523223] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1ee85e-8e34-4bb9-bc94-670455c7a79b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.606965] env[63418]: DEBUG oslo_concurrency.lockutils [req-209dbc94-ac15-4fda-8174-8affc9d4babf req-3e35b98a-2ded-4958-b52b-ff4f72c831f8 service nova] Releasing lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.683500] env[63418]: DEBUG oslo_concurrency.lockutils [req-b5b2b48e-38aa-47b5-b516-342f524bdc70 req-886cf0cc-b04c-46b0-b448-a3f30a296d7d service nova] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.693029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f4678410-5f61-4a1c-9e1d-657ae08f523a tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-0f99b32a-0125-4df0-919c-e5456b4ae4a5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.015s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.777681] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.777858] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquired lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.778040] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Forcefully refreshing network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 971.793488] env[63418]: DEBUG oslo_concurrency.lockutils [None req-dd79e803-85c2-4ce2-817d-a72f4b66446f tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "dcb52a03-8f82-4b33-8dc6-3924140db8d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.821s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.945018] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7596f511-fde2-41a3-af76-0ebd567a4571 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.953132] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64d6ebd-2416-4986-99aa-d95b34981759 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.990112] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74930b1b-3619-4ba9-93f4-2a044bf950a8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.001518] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d53795f-c6a2-4933-9f56-c15b3cf97dee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.005569] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245368, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538986} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.006015] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6/680c94e8-c6d9-4aab-92d4-047a0e70eec6.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 972.006396] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.007084] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f642997a-3446-405d-832a-d40841cc58b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.017037] env[63418]: DEBUG nova.compute.provider_tree [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.022649] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 972.022649] env[63418]: value = "task-1245370" [ 972.022649] env[63418]: _type = "Task" [ 972.022649] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.032293] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.034620] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.039270] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.039504] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1796e42-53b0-49fa-a8f5-e8f0b67f4b19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.054952] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 972.054952] env[63418]: value = "task-1245371" [ 972.054952] env[63418]: _type = "Task" [ 972.054952] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.063256] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.067946] env[63418]: DEBUG nova.network.neutron [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Successfully updated port: 050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.273434] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.273736] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquired lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.273834] env[63418]: DEBUG nova.network.neutron [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.295547] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 972.521354] env[63418]: DEBUG nova.scheduler.client.report [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 972.536981] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.564598] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245371, 'name': ReconfigVM_Task, 'duration_secs': 0.369687} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.564971] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 33 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.571860] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "refresh_cache-0c016d4e-ca34-4831-a567-e794012681db" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.572805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "refresh_cache-0c016d4e-ca34-4831-a567-e794012681db" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.572805] env[63418]: DEBUG nova.network.neutron [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.887660] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.934646] env[63418]: DEBUG nova.compute.manager [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Received event network-vif-plugged-050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 972.934881] env[63418]: DEBUG oslo_concurrency.lockutils [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] Acquiring lock "0c016d4e-ca34-4831-a567-e794012681db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.935133] env[63418]: DEBUG oslo_concurrency.lockutils [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] Lock "0c016d4e-ca34-4831-a567-e794012681db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.935284] env[63418]: DEBUG oslo_concurrency.lockutils [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] Lock "0c016d4e-ca34-4831-a567-e794012681db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.935481] env[63418]: DEBUG nova.compute.manager [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] No waiting events found dispatching network-vif-plugged-050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 972.935709] env[63418]: WARNING nova.compute.manager [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Received unexpected event network-vif-plugged-050c0715-5f26-4662-9135-1f933411cf3e for instance with vm_state building and task_state spawning. [ 972.935924] env[63418]: DEBUG nova.compute.manager [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Received event network-changed-050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 972.936137] env[63418]: DEBUG nova.compute.manager [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Refreshing instance network info cache due to event network-changed-050c0715-5f26-4662-9135-1f933411cf3e. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 972.936326] env[63418]: DEBUG oslo_concurrency.lockutils [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] Acquiring lock "refresh_cache-0c016d4e-ca34-4831-a567-e794012681db" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.035388] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.037622] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.643818} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.038711] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.223s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.038711] env[63418]: DEBUG nova.objects.instance [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'resources' on Instance uuid 7b0c70aa-e2bc-4131-97b4-4e53a378940a {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.039344] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 973.042249] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9613770-4992-4d5a-8932-e08e9bbea29b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.065784] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6/680c94e8-c6d9-4aab-92d4-047a0e70eec6.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.069238] env[63418]: INFO nova.scheduler.client.report [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance b0887bcd-7ba7-4c0e-8d50-886e27d37649 [ 973.070815] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b68b8ff7-bf86-4cfc-a024-5b8e012163bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.086988] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.087273] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.090015] env[63418]: DEBUG nova.virt.hardware [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.093976] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 973.100023] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a772c895-a2e5-4d3e-a68b-8ae3cf0c8582 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.120997] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 973.120997] env[63418]: value = "task-1245372" [ 973.120997] env[63418]: _type = "Task" [ 973.120997] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.125204] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 973.125204] env[63418]: value = "task-1245373" [ 973.125204] env[63418]: _type = "Task" [ 973.125204] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.135280] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245372, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.140454] env[63418]: DEBUG nova.network.neutron [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [{"id": "f786c95b-5214-454d-86f8-6d922f0482d8", "address": "fa:16:3e:ce:c4:76", "network": {"id": "819aaf6c-a126-497c-98f7-062f158ac742", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-486625765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0462b212fa4449c2a6f98cec2f186f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf786c95b-52", "ovs_interfaceid": "f786c95b-5214-454d-86f8-6d922f0482d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.140454] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245373, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.145375] env[63418]: DEBUG nova.network.neutron [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.299233] env[63418]: DEBUG nova.network.neutron [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Updating instance_info_cache with network_info: [{"id": "050c0715-5f26-4662-9135-1f933411cf3e", "address": "fa:16:3e:5e:68:f3", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap050c0715-5f", "ovs_interfaceid": "050c0715-5f26-4662-9135-1f933411cf3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.393915] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Releasing lock "refresh_cache-7b0c70aa-e2bc-4131-97b4-4e53a378940a" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.393915] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Updated the network info_cache for instance {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10349}} [ 973.393915] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.393915] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.393915] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.412663] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "interface-0f99b32a-0125-4df0-919c-e5456b4ae4a5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.413033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-0f99b32a-0125-4df0-919c-e5456b4ae4a5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.618753] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c3684755-35a2-48e2-9093-e5e9587a5986 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "b0887bcd-7ba7-4c0e-8d50-886e27d37649" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.722s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.637229] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245372, 'name': ReconfigVM_Task, 'duration_secs': 0.275922} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.643228] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6/680c94e8-c6d9-4aab-92d4-047a0e70eec6.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.644633] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Releasing lock "refresh_cache-6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.645247] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245373, 'name': ReconfigVM_Task, 'duration_secs': 0.17926} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.645368] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ebd417d-89d0-4eff-9ced-f678a04b42e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.647629] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54eda295-c488-4b3a-983c-df76072fefa3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.650075] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 973.651077] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2af4a8-82cf-4082-9928-3f5849f72b76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.659201] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Resuming the VM {{(pid=63418) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 973.659825] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 973.659825] env[63418]: value = "task-1245374" [ 973.659825] env[63418]: _type = "Task" [ 973.659825] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.670182] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-187ae10e-9645-470f-aad7-e8ce3075311b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.679548] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91/5dac16e3-06a9-443a-90c8-9aacdd23fd91.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.685645] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ae26399-4f34-4a3d-8b74-f1cb223322bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.704519] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245374, 'name': Rename_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.707572] env[63418]: DEBUG oslo_vmware.api [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 973.707572] env[63418]: value = "task-1245375" [ 973.707572] env[63418]: _type = "Task" [ 973.707572] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.707906] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 973.707906] env[63418]: value = "task-1245376" [ 973.707906] env[63418]: _type = "Task" [ 973.707906] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.722285] env[63418]: DEBUG oslo_vmware.api [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245375, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.725557] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245376, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.759053] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bcf759-b53b-43df-88a9-a39d60aa9e9c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.766513] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22617c7e-7633-478b-82e0-a2327f2294b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.798851] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1bfa77-8d7f-4ece-85ed-afb4801adc77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.801993] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "refresh_cache-0c016d4e-ca34-4831-a567-e794012681db" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.802309] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Instance network_info: |[{"id": "050c0715-5f26-4662-9135-1f933411cf3e", "address": "fa:16:3e:5e:68:f3", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap050c0715-5f", "ovs_interfaceid": "050c0715-5f26-4662-9135-1f933411cf3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 973.802618] env[63418]: DEBUG oslo_concurrency.lockutils [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] Acquired lock "refresh_cache-0c016d4e-ca34-4831-a567-e794012681db" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.802811] env[63418]: DEBUG nova.network.neutron [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Refreshing network info cache for port 050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.804605] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:68:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '050c0715-5f26-4662-9135-1f933411cf3e', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.812165] env[63418]: DEBUG oslo.service.loopingcall [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.813524] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.813761] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7f94659-37c8-44cc-92f9-36859f8dc7f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.831782] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e422ad1b-be42-4563-8748-9fdbfc9c1a8e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.836925] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.836925] env[63418]: value = "task-1245377" [ 973.836925] env[63418]: _type = "Task" [ 973.836925] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.848034] env[63418]: DEBUG nova.compute.provider_tree [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.853754] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245377, 'name': CreateVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.916056] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.916379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.917426] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff109d67-bf69-418e-afed-1969219fad2e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.935251] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe543a7-dec7-42fd-bb4b-6751aec062c6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.962900] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Reconfiguring VM to detach interface {{(pid=63418) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 973.963255] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e873de6-a145-45e9-9c7a-bb01e4c740e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.981754] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 973.981754] env[63418]: value = "task-1245378" [ 973.981754] env[63418]: _type = "Task" [ 973.981754] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.992475] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.182898] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245374, 'name': Rename_Task, 'duration_secs': 0.163585} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.182898] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.183125] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-399016c8-7090-4bf6-9e70-af686cf6a3b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.189453] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 974.189453] env[63418]: value = "task-1245379" [ 974.189453] env[63418]: _type = "Task" [ 974.189453] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.198830] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.221537] env[63418]: DEBUG oslo_vmware.api [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245375, 'name': PowerOnVM_Task} progress is 93%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.225130] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245376, 'name': ReconfigVM_Task, 'duration_secs': 0.395921} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.225481] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91/5dac16e3-06a9-443a-90c8-9aacdd23fd91.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.225849] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 50 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.313257] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.313611] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.348407] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245377, 'name': CreateVM_Task, 'duration_secs': 0.471019} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.351161] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.351945] env[63418]: DEBUG nova.scheduler.client.report [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 974.355632] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.355823] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.356196] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.357177] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea3d115c-f58b-4187-ac41-3d29ec9beef1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.363940] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 974.363940] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]524e7a89-ebdf-ab76-8347-4a1629473b77" [ 974.363940] env[63418]: _type = "Task" [ 974.363940] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.372126] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524e7a89-ebdf-ab76-8347-4a1629473b77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.409948] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.410224] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.493717] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.544877] env[63418]: DEBUG nova.network.neutron [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Updated VIF entry in instance network info cache for port 050c0715-5f26-4662-9135-1f933411cf3e. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.545429] env[63418]: DEBUG nova.network.neutron [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Updating instance_info_cache with network_info: [{"id": "050c0715-5f26-4662-9135-1f933411cf3e", "address": "fa:16:3e:5e:68:f3", "network": {"id": "08cb8137-c66d-4911-b40a-ea7cd565ea74", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1056041194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d445600834dd4c7e8022349ee993f3ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap050c0715-5f", "ovs_interfaceid": "050c0715-5f26-4662-9135-1f933411cf3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.699798] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245379, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.717968] env[63418]: DEBUG oslo_vmware.api [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245375, 'name': PowerOnVM_Task, 'duration_secs': 0.589591} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.718249] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Resumed the VM {{(pid=63418) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 974.718440] env[63418]: DEBUG nova.compute.manager [None req-2b2abdcc-b70b-4518-b3fd-f1a5c30b2083 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 974.719232] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7575cd3e-21a6-42c1-9610-c8d279f5fbe0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.733111] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b26bbe-4e12-4486-aa2b-20a8275f001b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.752243] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf7e584-eabb-48b3-bc47-e56e73cfc54a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.771208] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 67 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.819041] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 974.860068] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.861988] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.386s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.863641] env[63418]: INFO nova.compute.claims [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.876394] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]524e7a89-ebdf-ab76-8347-4a1629473b77, 'name': SearchDatastore_Task, 'duration_secs': 0.010634} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.876769] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.877091] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.877442] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.877619] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.877854] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.878161] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1682846e-5044-435f-9870-a76ac7f9de74 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.880786] env[63418]: INFO nova.scheduler.client.report [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted allocations for instance 7b0c70aa-e2bc-4131-97b4-4e53a378940a [ 974.889672] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.889962] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.890745] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55594f84-ecfe-47c6-8c49-3a1666c84985 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.897086] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 974.897086] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b1a357-4c7b-2b64-7fe2-1bc2905add57" [ 974.897086] env[63418]: _type = "Task" [ 974.897086] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.905893] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b1a357-4c7b-2b64-7fe2-1bc2905add57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.914148] env[63418]: DEBUG nova.compute.utils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.993346] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.048685] env[63418]: DEBUG oslo_concurrency.lockutils [req-da29d457-196a-44b7-9bde-0ad9b9eab2aa req-3c3432fd-053f-4453-9595-4d8b5b7557c4 service nova] Releasing lock "refresh_cache-0c016d4e-ca34-4831-a567-e794012681db" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.199844] env[63418]: DEBUG oslo_vmware.api [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245379, 'name': PowerOnVM_Task, 'duration_secs': 0.539603} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.200129] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.200342] env[63418]: INFO nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Took 8.43 seconds to spawn the instance on the hypervisor. [ 975.200528] env[63418]: DEBUG nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 975.201293] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2804ee8e-3ed8-4f90-81fb-a7fe97048763 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.313268] env[63418]: DEBUG nova.network.neutron [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Port 7ab35b58-105e-4976-b7f7-05e0c77de3fa binding to destination host cpu-1 is already ACTIVE {{(pid=63418) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 975.337936] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.390779] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a4e19779-6fc4-4712-8ff5-cf05450d8686 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "7b0c70aa-e2bc-4131-97b4-4e53a378940a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.958s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.407943] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52b1a357-4c7b-2b64-7fe2-1bc2905add57, 'name': SearchDatastore_Task, 'duration_secs': 0.022621} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.408864] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d073f989-818b-4d15-9493-025f9a3c4c34 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.414648] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 975.414648] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]520c09d5-34c2-3f97-0cea-5311ab9b5629" [ 975.414648] env[63418]: _type = "Task" [ 975.414648] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.418412] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.423672] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520c09d5-34c2-3f97-0cea-5311ab9b5629, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.494198] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.719921] env[63418]: INFO nova.compute.manager [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Took 13.34 seconds to build instance. [ 975.926843] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]520c09d5-34c2-3f97-0cea-5311ab9b5629, 'name': SearchDatastore_Task, 'duration_secs': 0.009401} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.927859] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.928149] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0c016d4e-ca34-4831-a567-e794012681db/0c016d4e-ca34-4831-a567-e794012681db.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.928456] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be4472c7-6d4d-4d78-9915-22501bae937d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.935225] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 975.935225] env[63418]: value = "task-1245380" [ 975.935225] env[63418]: _type = "Task" [ 975.935225] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.946391] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245380, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.999965] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.052974] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03996dd-0d1a-4811-b2b8-3df59b80d66a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.060616] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55306913-5d54-4592-9b4a-13066fcea1d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.091927] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030a8dad-4e46-4f22-9371-4eb7cd9ad4ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.099359] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0674ed8-f954-47e9-ad64-0ef0532fb996 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.113221] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.222054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f20f580a-6cef-4e79-b881-ab1996be2d37 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.856s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.337799] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.337799] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.337799] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.448017] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245380, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464399} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.448017] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 0c016d4e-ca34-4831-a567-e794012681db/0c016d4e-ca34-4831-a567-e794012681db.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.448017] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.448017] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5fe2e0f-6a00-426d-b285-87332f8b2cea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.453583] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 976.453583] env[63418]: value = "task-1245381" [ 976.453583] env[63418]: _type = "Task" [ 976.453583] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.461542] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245381, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.488026] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.488352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.488716] env[63418]: INFO nova.compute.manager [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Attaching volume e5151ce5-c6ee-4092-9eeb-e00affc031dd to /dev/sdb [ 976.507744] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.523159] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.523159] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.533697] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c1ab65-69a7-4b7c-81d8-e5942bf39d5b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.543400] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702ee6bc-4f4c-4607-ac69-1e6a34fb0f21 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.558679] env[63418]: DEBUG nova.virt.block_device [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating existing volume attachment record: 96944125-b41a-41e5-9166-d0fdfb65e6dc {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 976.639815] env[63418]: ERROR nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [req-0246bb62-8ad8-46e6-a48a-4d0c680eebf0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0246bb62-8ad8-46e6-a48a-4d0c680eebf0"}]} [ 976.656938] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 976.673068] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 976.673332] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.689518] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 976.710789] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 976.883973] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f974610-a5d3-489e-b86e-553a8f06555e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.897724] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c7b0f2-6bab-4f38-916c-cbf66e31e3fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.929929] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccc70a9-9b8b-41a3-a47e-57e6f8204955 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.938304] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9579132d-d4b6-4567-bb01-0ba64f99ab74 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.952229] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.963707] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245381, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063492} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.964593] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.965383] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add14eab-ee76-4014-95c6-f9169313ada4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.988063] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 0c016d4e-ca34-4831-a567-e794012681db/0c016d4e-ca34-4831-a567-e794012681db.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.988745] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e980ec4a-85ca-42d7-ab98-5fcdd7da1f50 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.011342] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.012715] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 977.012715] env[63418]: value = "task-1245385" [ 977.012715] env[63418]: _type = "Task" [ 977.012715] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.020219] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245385, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.025736] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 977.375675] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.375972] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.376197] env[63418]: DEBUG nova.network.neutron [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.476801] env[63418]: ERROR nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [req-9bcffa76-dc56-45ef-926c-39fc59dc78dc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9bcffa76-dc56-45ef-926c-39fc59dc78dc"}]} [ 977.493685] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 977.501292] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.506374] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 977.506665] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.518098] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 977.525277] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245385, 'name': ReconfigVM_Task, 'duration_secs': 0.430122} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.525763] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 0c016d4e-ca34-4831-a567-e794012681db/0c016d4e-ca34-4831-a567-e794012681db.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.526449] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e80c8ea-f14f-4184-aa30-f3fd58f3c05e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.535091] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 977.535091] env[63418]: value = "task-1245386" [ 977.535091] env[63418]: _type = "Task" [ 977.535091] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.535985] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 977.549104] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245386, 'name': Rename_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.552130] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.621106] env[63418]: DEBUG nova.compute.manager [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Stashing vm_state: active {{(pid=63418) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 977.725031] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b210428-3386-42f5-98f1-961efa5baee8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.732883] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80f1e87-6cf4-4bac-bd16-ed101724faad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.765697] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed18a64-ad67-40ab-b6ef-e1ed07b642e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.776044] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd6d60a-b372-443b-bad2-1abce860a91f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.792865] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.999829] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.049474] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245386, 'name': Rename_Task, 'duration_secs': 0.145972} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.049831] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.050148] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44da4c9c-8f28-4212-b5a4-342458388b0e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.056623] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 978.056623] env[63418]: value = "task-1245387" [ 978.056623] env[63418]: _type = "Task" [ 978.056623] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.065139] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245387, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.070819] env[63418]: DEBUG nova.network.neutron [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [{"id": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "address": "fa:16:3e:43:d6:a0", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ab35b58-10", "ovs_interfaceid": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.137966] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.325138] env[63418]: DEBUG nova.scheduler.client.report [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 978.325138] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 123 to 124 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 978.325138] env[63418]: DEBUG nova.compute.provider_tree [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 978.499334] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.566609] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245387, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.573311] env[63418]: DEBUG oslo_concurrency.lockutils [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.830350] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.969s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.831308] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 978.834061] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.496s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.835518] env[63418]: INFO nova.compute.claims [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.000424] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.070028] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245387, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.100838] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7d80e9-2943-4737-999a-eaf06548674c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.121073] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f46a4b1-2b1b-4df4-8c06-900da9535c77 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.128823] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 83 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 979.340114] env[63418]: DEBUG nova.compute.utils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.343090] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 979.343267] env[63418]: DEBUG nova.network.neutron [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.383380] env[63418]: DEBUG nova.policy [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b29941866349482fb9e53dcf87cb1845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05ef1c6c74574217817c6ab14a022b91', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 979.500606] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.569906] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245387, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.635717] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.636050] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2444b311-5df2-40c2-8ee6-15f61295a83c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.643197] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 979.643197] env[63418]: value = "task-1245389" [ 979.643197] env[63418]: _type = "Task" [ 979.643197] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.651601] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.671495] env[63418]: DEBUG nova.network.neutron [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Successfully created port: a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.844885] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 980.003394] env[63418]: DEBUG oslo_vmware.api [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245378, 'name': ReconfigVM_Task, 'duration_secs': 5.951658} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.005021] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.005021] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Reconfigured VM to detach interface {{(pid=63418) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 980.052622] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746b560b-d69e-43ab-95d9-dba0fbab8788 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.066294] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48267998-b8f0-4015-839f-dc3aca6605d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.074020] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245387, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.099620] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4bd3d1e-c918-498d-9fde-36ee020619f5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.107363] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d65a49-f5a4-4a0a-b6bf-41f882fcff4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.121982] env[63418]: DEBUG nova.compute.provider_tree [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.152806] env[63418]: DEBUG oslo_vmware.api [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245389, 'name': PowerOnVM_Task, 'duration_secs': 0.351531} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.153146] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.153322] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-ee470cc0-dd40-4f49-a452-ced505c8470b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance '5dac16e3-06a9-443a-90c8-9aacdd23fd91' progress to 100 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 980.477381] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.477659] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.477880] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.478085] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.478265] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.480898] env[63418]: INFO nova.compute.manager [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Terminating instance [ 980.569906] env[63418]: DEBUG oslo_vmware.api [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245387, 'name': PowerOnVM_Task, 'duration_secs': 2.403757} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.570267] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.570385] env[63418]: INFO nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Took 9.09 seconds to spawn the instance on the hypervisor. [ 980.570566] env[63418]: DEBUG nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 980.571327] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6909b2a8-291f-4b48-a1db-956180337e95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.626948] env[63418]: DEBUG nova.scheduler.client.report [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 980.859808] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 980.885653] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.885952] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.886155] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.886349] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.886502] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.886653] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.886925] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.887116] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.887328] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.887522] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.887699] env[63418]: DEBUG nova.virt.hardware [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.888635] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8cd800-f309-4e8c-b93c-b79210bc59f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.896187] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d087eeb-8b7d-4781-8d6b-f53cd8b6f662 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.985091] env[63418]: DEBUG nova.compute.manager [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 980.985350] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.986288] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8104647-6d2c-4200-9052-89613e6abc7b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.993724] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.993963] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5b8692c-692e-4747-b245-c5cb9b434af3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.000289] env[63418]: DEBUG oslo_vmware.api [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 981.000289] env[63418]: value = "task-1245390" [ 981.000289] env[63418]: _type = "Task" [ 981.000289] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.008583] env[63418]: DEBUG oslo_vmware.api [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.087396] env[63418]: INFO nova.compute.manager [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Took 16.10 seconds to build instance. [ 981.090639] env[63418]: DEBUG nova.compute.manager [req-cab88107-c4cd-4338-a090-1d4770d1de82 req-53f63605-f9cb-4f70-b390-f227cde41fe1 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Received event network-vif-plugged-a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 981.090867] env[63418]: DEBUG oslo_concurrency.lockutils [req-cab88107-c4cd-4338-a090-1d4770d1de82 req-53f63605-f9cb-4f70-b390-f227cde41fe1 service nova] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.091145] env[63418]: DEBUG oslo_concurrency.lockutils [req-cab88107-c4cd-4338-a090-1d4770d1de82 req-53f63605-f9cb-4f70-b390-f227cde41fe1 service nova] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.091331] env[63418]: DEBUG oslo_concurrency.lockutils [req-cab88107-c4cd-4338-a090-1d4770d1de82 req-53f63605-f9cb-4f70-b390-f227cde41fe1 service nova] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.091525] env[63418]: DEBUG nova.compute.manager [req-cab88107-c4cd-4338-a090-1d4770d1de82 req-53f63605-f9cb-4f70-b390-f227cde41fe1 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] No waiting events found dispatching network-vif-plugged-a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 981.091713] env[63418]: WARNING nova.compute.manager [req-cab88107-c4cd-4338-a090-1d4770d1de82 req-53f63605-f9cb-4f70-b390-f227cde41fe1 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Received unexpected event network-vif-plugged-a6337c68-709c-4b53-b00a-b0e794cd45db for instance with vm_state building and task_state spawning. [ 981.114979] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 981.115346] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268519', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'name': 'volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcbaf4f-ee24-4072-83ae-ffde59478928', 'attached_at': '', 'detached_at': '', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'serial': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 981.116334] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7160fd-c04c-4138-9fdc-e3a7275e4ce4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.133531] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.134061] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 981.137580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.585s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.138361] env[63418]: INFO nova.compute.claims [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.141943] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6ce0fd-0c7f-4702-a1bd-a9467ee73136 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.173708] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd/volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.174921] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d382b37-ddb2-46c6-81cf-012809effdc1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.194609] env[63418]: DEBUG oslo_vmware.api [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 981.194609] env[63418]: value = "task-1245391" [ 981.194609] env[63418]: _type = "Task" [ 981.194609] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.204345] env[63418]: DEBUG oslo_vmware.api [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245391, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.205248] env[63418]: DEBUG nova.network.neutron [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Successfully updated port: a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.509890] env[63418]: DEBUG oslo_vmware.api [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245390, 'name': PowerOffVM_Task, 'duration_secs': 0.278796} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.510284] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.510541] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.510785] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c598ea95-0498-4c39-9a64-f59dc9cb8222 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.519718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.519981] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquired lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.520262] env[63418]: DEBUG nova.network.neutron [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.593075] env[63418]: DEBUG oslo_concurrency.lockutils [None req-af9180a6-e602-4d11-aeaf-9e59dfb9bbca tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0c016d4e-ca34-4831-a567-e794012681db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.610s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.619266] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.619499] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.619690] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleting the datastore file [datastore1] 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.619942] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da1ae723-a152-4ac3-9593-40b24c66d75d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.626650] env[63418]: DEBUG oslo_vmware.api [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for the task: (returnval){ [ 981.626650] env[63418]: value = "task-1245393" [ 981.626650] env[63418]: _type = "Task" [ 981.626650] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.636031] env[63418]: DEBUG oslo_vmware.api [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245393, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.647855] env[63418]: DEBUG nova.compute.utils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.651576] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 981.651777] env[63418]: DEBUG nova.network.neutron [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.702834] env[63418]: DEBUG nova.policy [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.709595] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.709768] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.709926] env[63418]: DEBUG nova.network.neutron [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.711534] env[63418]: DEBUG oslo_vmware.api [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245391, 'name': ReconfigVM_Task, 'duration_secs': 0.32622} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.712034] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Reconfigured VM instance instance-00000058 to attach disk [datastore2] volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd/volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.719394] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5cc1dde-8cfb-4e30-8e46-d52538462ee6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.736286] env[63418]: DEBUG oslo_vmware.api [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 981.736286] env[63418]: value = "task-1245394" [ 981.736286] env[63418]: _type = "Task" [ 981.736286] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.746456] env[63418]: DEBUG oslo_vmware.api [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245394, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.979335] env[63418]: DEBUG nova.network.neutron [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Successfully created port: 7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.982161] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.982562] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.982890] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.983228] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.983524] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.987973] env[63418]: INFO nova.compute.manager [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Terminating instance [ 982.118153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.118153] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.118153] env[63418]: DEBUG nova.compute.manager [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Going to confirm migration 3 {{(pid=63418) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 982.141306] env[63418]: DEBUG oslo_vmware.api [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Task: {'id': task-1245393, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14047} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.141306] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.141306] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.141306] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.141306] env[63418]: INFO nova.compute.manager [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 982.141306] env[63418]: DEBUG oslo.service.loopingcall [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.141306] env[63418]: DEBUG nova.compute.manager [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 982.141306] env[63418]: DEBUG nova.network.neutron [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 982.155151] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 982.250021] env[63418]: DEBUG oslo_vmware.api [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245394, 'name': ReconfigVM_Task, 'duration_secs': 0.130308} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.252932] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268519', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'name': 'volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcbaf4f-ee24-4072-83ae-ffde59478928', 'attached_at': '', 'detached_at': '', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'serial': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 982.266962] env[63418]: DEBUG nova.network.neutron [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.385895] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e6919c-f08d-4231-a8a7-5939e2045c4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.393566] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2534385f-2b0e-4e50-9da6-313effbf6da9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.426402] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e939cc4-67e7-446a-9712-90f04d77bc8d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.433970] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d08237-06b4-48ad-af92-bd98f813ee47 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.438687] env[63418]: INFO nova.network.neutron [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Port f1f171f0-8f67-4b5b-a9c8-91014404b4e2 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 982.439447] env[63418]: DEBUG nova.network.neutron [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [{"id": "34d14910-009f-4ee8-b718-a43961c430cb", "address": "fa:16:3e:96:7f:a4", "network": {"id": "f11c9d94-8e8f-4c7d-b518-f8958080b8fa", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-656736173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4297b53faeab40dfa5de863ad4030800", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d14910-00", "ovs_interfaceid": "34d14910-009f-4ee8-b718-a43961c430cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.456020] env[63418]: DEBUG nova.compute.provider_tree [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.493703] env[63418]: DEBUG nova.compute.manager [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 982.493802] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.495740] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cef09d6-5401-486d-935a-510d0c4bf4a5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.507295] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.507594] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfe4bd0f-fc08-4abb-b3f6-3d5835130f09 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.514540] env[63418]: DEBUG oslo_vmware.api [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 982.514540] env[63418]: value = "task-1245395" [ 982.514540] env[63418]: _type = "Task" [ 982.514540] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.522423] env[63418]: DEBUG oslo_vmware.api [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.526668] env[63418]: DEBUG nova.compute.manager [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 982.527562] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febebd4b-7b33-4eb9-b3f4-3875b84d2e4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.569023] env[63418]: DEBUG nova.network.neutron [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating instance_info_cache with network_info: [{"id": "a6337c68-709c-4b53-b00a-b0e794cd45db", "address": "fa:16:3e:78:ae:27", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6337c68-70", "ovs_interfaceid": "a6337c68-709c-4b53-b00a-b0e794cd45db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.699274] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.699537] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.699668] env[63418]: DEBUG nova.network.neutron [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.699863] env[63418]: DEBUG nova.objects.instance [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lazy-loading 'info_cache' on Instance uuid 5dac16e3-06a9-443a-90c8-9aacdd23fd91 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.943065] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Releasing lock "refresh_cache-0f99b32a-0125-4df0-919c-e5456b4ae4a5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.987776] env[63418]: DEBUG nova.scheduler.client.report [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 124 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 982.988125] env[63418]: DEBUG nova.compute.provider_tree [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 124 to 125 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 982.988328] env[63418]: DEBUG nova.compute.provider_tree [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 983.026419] env[63418]: DEBUG oslo_vmware.api [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245395, 'name': PowerOffVM_Task, 'duration_secs': 0.232947} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.026930] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.027252] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.027636] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0eadd7e-96f2-4ba8-a946-1b1fbe59a49f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.037417] env[63418]: INFO nova.compute.manager [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] instance snapshotting [ 983.039973] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29497ed2-6155-4ead-992f-0e23e3cc5669 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.060869] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e60987-ee0c-45ab-b5f5-2cf921a64c73 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.072286] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.072627] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Instance network_info: |[{"id": "a6337c68-709c-4b53-b00a-b0e794cd45db", "address": "fa:16:3e:78:ae:27", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6337c68-70", "ovs_interfaceid": "a6337c68-709c-4b53-b00a-b0e794cd45db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 983.073237] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:ae:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'af454577-0e89-41a3-a9f2-f39716f62fd5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6337c68-709c-4b53-b00a-b0e794cd45db', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.081253] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Creating folder: Project (05ef1c6c74574217817c6ab14a022b91). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 983.081534] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f1362ea-43d4-4f72-a586-6b798e595094 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.092662] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Created folder: Project (05ef1c6c74574217817c6ab14a022b91) in parent group-v268354. [ 983.092898] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Creating folder: Instances. Parent ref: group-v268520. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 983.093158] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4af37c51-9faf-4534-a5b5-2ecfcb1b4ea9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.104833] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Created folder: Instances in parent group-v268520. [ 983.105118] env[63418]: DEBUG oslo.service.loopingcall [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.105317] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 983.105820] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca9307d5-03e9-494b-a048-1d69b86f5c99 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.122610] env[63418]: DEBUG nova.network.neutron [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.128132] env[63418]: DEBUG nova.compute.manager [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Received event network-changed-a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 983.128132] env[63418]: DEBUG nova.compute.manager [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Refreshing instance network info cache due to event network-changed-a6337c68-709c-4b53-b00a-b0e794cd45db. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 983.128132] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] Acquiring lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.128132] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] Acquired lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.128132] env[63418]: DEBUG nova.network.neutron [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Refreshing network info cache for port a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.128132] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.128450] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.128450] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleting the datastore file [datastore2] 0f99b32a-0125-4df0-919c-e5456b4ae4a5 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.129400] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8c37cfc-20f4-4849-ba49-c01591d93542 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.134075] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.134075] env[63418]: value = "task-1245399" [ 983.134075] env[63418]: _type = "Task" [ 983.134075] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.139630] env[63418]: DEBUG oslo_vmware.api [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 983.139630] env[63418]: value = "task-1245400" [ 983.139630] env[63418]: _type = "Task" [ 983.139630] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.146410] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245399, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.152041] env[63418]: DEBUG oslo_vmware.api [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.173122] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 983.201139] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.201554] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.201811] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.202138] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.202418] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.202649] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.202970] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.203245] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.203505] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.203775] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.204072] env[63418]: DEBUG nova.virt.hardware [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.207259] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7a9e65-84f8-4946-8334-f623fb130507 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.220217] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887d7538-30f0-4edf-a4df-7302ff91d023 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.299332] env[63418]: DEBUG nova.objects.instance [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.447221] env[63418]: DEBUG oslo_concurrency.lockutils [None req-26b1e5f0-fe1c-4bd3-913e-ae76cb5bf6e0 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "interface-0f99b32a-0125-4df0-919c-e5456b4ae4a5-f1f171f0-8f67-4b5b-a9c8-91014404b4e2" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.034s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.496018] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.496018] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 983.498554] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.361s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.574451] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 983.575790] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-70298382-e9b8-49af-ab37-7eb7914f5e80 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.584205] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 983.584205] env[63418]: value = "task-1245401" [ 983.584205] env[63418]: _type = "Task" [ 983.584205] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.592846] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245401, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.629866] env[63418]: INFO nova.compute.manager [-] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Took 1.49 seconds to deallocate network for instance. [ 983.655076] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245399, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.659401] env[63418]: DEBUG oslo_vmware.api [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153503} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.659724] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.659981] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.660566] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.660843] env[63418]: INFO nova.compute.manager [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Took 1.17 seconds to destroy the instance on the hypervisor. [ 983.661295] env[63418]: DEBUG oslo.service.loopingcall [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.661580] env[63418]: DEBUG nova.compute.manager [-] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 983.661723] env[63418]: DEBUG nova.network.neutron [-] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 983.804878] env[63418]: DEBUG oslo_concurrency.lockutils [None req-bf5844f9-8589-4b57-bfac-955c978c30cc tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.316s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.837914] env[63418]: DEBUG nova.network.neutron [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Successfully updated port: 7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.891889] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.892580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.893038] env[63418]: DEBUG nova.compute.manager [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 983.894082] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f147f57d-78db-4709-b959-7bf3f259b89a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.908297] env[63418]: DEBUG nova.compute.manager [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 983.909121] env[63418]: DEBUG nova.objects.instance [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.980483] env[63418]: DEBUG nova.network.neutron [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [{"id": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "address": "fa:16:3e:43:d6:a0", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ab35b58-10", "ovs_interfaceid": "7ab35b58-105e-4976-b7f7-05e0c77de3fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.000410] env[63418]: DEBUG nova.compute.utils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.005902] env[63418]: INFO nova.compute.claims [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.009240] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 984.009402] env[63418]: DEBUG nova.network.neutron [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.023803] env[63418]: DEBUG nova.network.neutron [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updated VIF entry in instance network info cache for port a6337c68-709c-4b53-b00a-b0e794cd45db. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.025376] env[63418]: DEBUG nova.network.neutron [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating instance_info_cache with network_info: [{"id": "a6337c68-709c-4b53-b00a-b0e794cd45db", "address": "fa:16:3e:78:ae:27", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6337c68-70", "ovs_interfaceid": "a6337c68-709c-4b53-b00a-b0e794cd45db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.058271] env[63418]: DEBUG nova.policy [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e9d38547e714839a808ae0c7e7e97d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '381605706a8b4a1e9cca8603c800f1ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.094660] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245401, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.146129] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.146400] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245399, 'name': CreateVM_Task, 'duration_secs': 0.629442} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.146556] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 984.147247] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.147418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.147749] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.148015] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab9bf3dc-9ecf-4f0c-afd9-2c4cba076793 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.152995] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 984.152995] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5247982b-ae19-8735-8780-dddb5615cf8b" [ 984.152995] env[63418]: _type = "Task" [ 984.152995] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.163549] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5247982b-ae19-8735-8780-dddb5615cf8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.339326] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-3e645774-d1b3-468b-86a5-ff0c07e77f4e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.339518] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-3e645774-d1b3-468b-86a5-ff0c07e77f4e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.339684] env[63418]: DEBUG nova.network.neutron [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.484816] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-5dac16e3-06a9-443a-90c8-9aacdd23fd91" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.484816] env[63418]: DEBUG nova.objects.instance [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lazy-loading 'migration_context' on Instance uuid 5dac16e3-06a9-443a-90c8-9aacdd23fd91 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.512303] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 984.519134] env[63418]: INFO nova.compute.resource_tracker [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating resource usage from migration 1a492728-e2b4-48fb-bec5-473405172064 [ 984.528920] env[63418]: DEBUG oslo_concurrency.lockutils [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] Releasing lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.528920] env[63418]: DEBUG nova.compute.manager [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Received event network-vif-deleted-f786c95b-5214-454d-86f8-6d922f0482d8 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 984.528920] env[63418]: INFO nova.compute.manager [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Neutron deleted interface f786c95b-5214-454d-86f8-6d922f0482d8; detaching it from the instance and deleting it from the info cache [ 984.528920] env[63418]: DEBUG nova.network.neutron [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.595054] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245401, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.668453] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5247982b-ae19-8735-8780-dddb5615cf8b, 'name': SearchDatastore_Task, 'duration_secs': 0.012338} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.668453] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.668453] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.668729] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.669396] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.669554] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.672566] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c6d85a6-e04f-4c33-85e0-c788a05e573e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.683123] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.683123] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 984.683123] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72af4c20-2985-45c1-acf0-93e92abf494f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.687456] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 984.687456] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a55044-f85e-e2f7-1488-b0944f511363" [ 984.687456] env[63418]: _type = "Task" [ 984.687456] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.700573] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a55044-f85e-e2f7-1488-b0944f511363, 'name': SearchDatastore_Task, 'duration_secs': 0.007308} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.703885] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0e4c2a3-9dd5-4795-b6dc-0787dbd535cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.706600] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 984.706600] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52764310-90e9-8222-1132-f7cb3136743c" [ 984.706600] env[63418]: _type = "Task" [ 984.706600] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.723160] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52764310-90e9-8222-1132-f7cb3136743c, 'name': SearchDatastore_Task, 'duration_secs': 0.007433} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.723160] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.723330] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 98f51fcd-951c-4c16-bc20-efc62ef359f5/98f51fcd-951c-4c16-bc20-efc62ef359f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.723609] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54172036-2fcf-4cb2-bf9b-0138e1b7aaa4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.731040] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 984.731040] env[63418]: value = "task-1245402" [ 984.731040] env[63418]: _type = "Task" [ 984.731040] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.744519] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245402, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.770820] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfecd92-4c4e-48c5-8300-dc63fec08b2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.777027] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f592b1e-ebde-43e4-8d33-4a2b9e1ea061 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.780910] env[63418]: DEBUG nova.network.neutron [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Successfully created port: 6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.817822] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa334f03-b5fa-4925-9887-61e4dade2810 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.825532] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda5b393-1b31-42a7-b40a-09d3d4bfabd8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.839138] env[63418]: DEBUG nova.compute.provider_tree [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.871964] env[63418]: DEBUG nova.network.neutron [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.916503] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.916842] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d63c90ea-6510-484c-8ae0-69582af4e053 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.925158] env[63418]: DEBUG oslo_vmware.api [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 984.925158] env[63418]: value = "task-1245403" [ 984.925158] env[63418]: _type = "Task" [ 984.925158] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.934356] env[63418]: DEBUG oslo_vmware.api [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.988571] env[63418]: DEBUG nova.objects.base [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Object Instance<5dac16e3-06a9-443a-90c8-9aacdd23fd91> lazy-loaded attributes: info_cache,migration_context {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 984.989912] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c086777-6b8c-4d84-a429-6cb096932f18 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.993984] env[63418]: DEBUG nova.network.neutron [-] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.013639] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-482d7891-c3b1-4639-87e6-c372a8c562f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.020778] env[63418]: DEBUG oslo_vmware.api [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 985.020778] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5211c9ea-2241-873a-78d3-a96ea4f09cc6" [ 985.020778] env[63418]: _type = "Task" [ 985.020778] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.031661] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-650d2b0b-590b-471d-988f-e999365c54c9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.039861] env[63418]: DEBUG oslo_vmware.api [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5211c9ea-2241-873a-78d3-a96ea4f09cc6, 'name': SearchDatastore_Task, 'duration_secs': 0.014945} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.041882] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.046298] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b4747d-cce5-414c-9c39-a458bde87bce {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.056366] env[63418]: DEBUG nova.network.neutron [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Updating instance_info_cache with network_info: [{"id": "7281180f-8ed9-4bba-a220-c22a98d334c7", "address": "fa:16:3e:43:96:32", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7281180f-8e", "ovs_interfaceid": "7281180f-8ed9-4bba-a220-c22a98d334c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.076318] env[63418]: DEBUG nova.compute.manager [req-9ff8a8ce-da9f-456c-aa29-df6c4c1dd5b5 req-ff0fb0b7-865f-4760-a791-f0b63e0f1b61 service nova] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Detach interface failed, port_id=f786c95b-5214-454d-86f8-6d922f0482d8, reason: Instance 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 985.096655] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245401, 'name': CreateSnapshot_Task, 'duration_secs': 1.210857} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.097020] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 985.097883] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714bd587-e47c-433b-870a-00c203737a6e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.158470] env[63418]: DEBUG nova.compute.manager [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Received event network-vif-plugged-7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 985.158720] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] Acquiring lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.158968] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.159271] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.159530] env[63418]: DEBUG nova.compute.manager [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] No waiting events found dispatching network-vif-plugged-7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 985.159730] env[63418]: WARNING nova.compute.manager [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Received unexpected event network-vif-plugged-7281180f-8ed9-4bba-a220-c22a98d334c7 for instance with vm_state building and task_state spawning. [ 985.159878] env[63418]: DEBUG nova.compute.manager [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Received event network-changed-7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 985.160072] env[63418]: DEBUG nova.compute.manager [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Refreshing instance network info cache due to event network-changed-7281180f-8ed9-4bba-a220-c22a98d334c7. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 985.160316] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] Acquiring lock "refresh_cache-3e645774-d1b3-468b-86a5-ff0c07e77f4e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.241269] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245402, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479232} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.241584] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 98f51fcd-951c-4c16-bc20-efc62ef359f5/98f51fcd-951c-4c16-bc20-efc62ef359f5.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.241812] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.242100] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab89c222-8d7b-45ce-984e-a1a832d7ffba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.248781] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 985.248781] env[63418]: value = "task-1245404" [ 985.248781] env[63418]: _type = "Task" [ 985.248781] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.257458] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245404, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.341834] env[63418]: DEBUG nova.scheduler.client.report [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 985.435805] env[63418]: DEBUG oslo_vmware.api [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245403, 'name': PowerOffVM_Task, 'duration_secs': 0.410684} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.436100] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.436308] env[63418]: DEBUG nova.compute.manager [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 985.437124] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b33a59-ff81-4f66-a147-3170b942ae6a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.497569] env[63418]: INFO nova.compute.manager [-] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Took 1.84 seconds to deallocate network for instance. [ 985.533307] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 985.558645] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-3e645774-d1b3-468b-86a5-ff0c07e77f4e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.559070] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Instance network_info: |[{"id": "7281180f-8ed9-4bba-a220-c22a98d334c7", "address": "fa:16:3e:43:96:32", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7281180f-8e", "ovs_interfaceid": "7281180f-8ed9-4bba-a220-c22a98d334c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 985.561201] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.561447] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.561826] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.561826] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.561956] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.562133] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.562284] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.562604] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.562814] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.562984] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.563178] env[63418]: DEBUG nova.virt.hardware [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.563468] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] Acquired lock "refresh_cache-3e645774-d1b3-468b-86a5-ff0c07e77f4e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.563669] env[63418]: DEBUG nova.network.neutron [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Refreshing network info cache for port 7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.564813] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:96:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7281180f-8ed9-4bba-a220-c22a98d334c7', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.572288] env[63418]: DEBUG oslo.service.loopingcall [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.573648] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d3e5a4-138a-4361-8e4e-3a1725210e4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.578461] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.578944] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8198fb93-3773-4c6a-a712-582db626c589 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.598861] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccf3eb4-0901-4b4a-b461-e05336f182cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.604129] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.604129] env[63418]: value = "task-1245405" [ 985.604129] env[63418]: _type = "Task" [ 985.604129] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.621388] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 985.622521] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b1fc3690-3baa-4d04-9f3b-7d627e9829f3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.628069] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245405, 'name': CreateVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.634903] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 985.634903] env[63418]: value = "task-1245406" [ 985.634903] env[63418]: _type = "Task" [ 985.634903] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.642726] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245406, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.758881] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245404, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.249005} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.759175] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.759988] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf14b8b-7fca-48a8-9d4c-d5d312b0d5bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.782768] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 98f51fcd-951c-4c16-bc20-efc62ef359f5/98f51fcd-951c-4c16-bc20-efc62ef359f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.785401] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf84ac58-fa43-48c5-a1fb-7df840ebedcf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.805328] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 985.805328] env[63418]: value = "task-1245407" [ 985.805328] env[63418]: _type = "Task" [ 985.805328] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.813390] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.838835] env[63418]: DEBUG nova.network.neutron [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Updated VIF entry in instance network info cache for port 7281180f-8ed9-4bba-a220-c22a98d334c7. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 985.839233] env[63418]: DEBUG nova.network.neutron [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Updating instance_info_cache with network_info: [{"id": "7281180f-8ed9-4bba-a220-c22a98d334c7", "address": "fa:16:3e:43:96:32", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7281180f-8e", "ovs_interfaceid": "7281180f-8ed9-4bba-a220-c22a98d334c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.848612] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.350s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.848816] env[63418]: INFO nova.compute.manager [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Migrating [ 985.855805] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.710s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.856307] env[63418]: DEBUG nova.objects.instance [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lazy-loading 'resources' on Instance uuid 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.949282] env[63418]: DEBUG oslo_concurrency.lockutils [None req-8f577da3-ac5e-41bd-a0fb-27a74dc3104c tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.057s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.008603] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.114904] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245405, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.145634] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245406, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.315645] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.341450] env[63418]: DEBUG oslo_concurrency.lockutils [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] Releasing lock "refresh_cache-3e645774-d1b3-468b-86a5-ff0c07e77f4e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.341738] env[63418]: DEBUG nova.compute.manager [req-7d125113-a6cb-4019-8a0f-55da3f345c72 req-77b4b52f-fff8-4c0e-a6b6-2c88b98d0f4a service nova] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Received event network-vif-deleted-34d14910-009f-4ee8-b718-a43961c430cb {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 986.356543] env[63418]: DEBUG nova.compute.manager [req-672754ba-7972-4eae-8a7b-de569dddf21c req-4431e67e-c7a5-44bb-9c68-322291daf77c service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Received event network-vif-plugged-6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 986.356767] env[63418]: DEBUG oslo_concurrency.lockutils [req-672754ba-7972-4eae-8a7b-de569dddf21c req-4431e67e-c7a5-44bb-9c68-322291daf77c service nova] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.357018] env[63418]: DEBUG oslo_concurrency.lockutils [req-672754ba-7972-4eae-8a7b-de569dddf21c req-4431e67e-c7a5-44bb-9c68-322291daf77c service nova] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.357266] env[63418]: DEBUG oslo_concurrency.lockutils [req-672754ba-7972-4eae-8a7b-de569dddf21c req-4431e67e-c7a5-44bb-9c68-322291daf77c service nova] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.357449] env[63418]: DEBUG nova.compute.manager [req-672754ba-7972-4eae-8a7b-de569dddf21c req-4431e67e-c7a5-44bb-9c68-322291daf77c service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] No waiting events found dispatching network-vif-plugged-6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 986.357620] env[63418]: WARNING nova.compute.manager [req-672754ba-7972-4eae-8a7b-de569dddf21c req-4431e67e-c7a5-44bb-9c68-322291daf77c service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Received unexpected event network-vif-plugged-6eb8c1f9-eec0-4763-8033-34a4efda65c1 for instance with vm_state building and task_state spawning. [ 986.368551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.368551] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.368551] env[63418]: DEBUG nova.network.neutron [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.449157] env[63418]: DEBUG nova.network.neutron [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Successfully updated port: 6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.456093] env[63418]: DEBUG nova.objects.instance [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.573517] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a351fcf3-d059-498d-a1b6-89ec2c4a6b10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.581201] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400bc43a-547c-49d4-9aeb-431bbb5732a1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.613770] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c547981e-1937-4541-8770-affaeca5d3c5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.622579] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245405, 'name': CreateVM_Task, 'duration_secs': 0.682591} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.624506] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.625268] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.625438] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.625783] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.626976] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3b231f-f103-41f7-8be7-89aa7a7cffcd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.630619] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce3d65ce-9c5d-4b73-b221-e9b1048e36aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.641650] env[63418]: DEBUG nova.compute.provider_tree [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.646468] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 986.646468] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5244102b-a988-5842-ced1-e0f574599f2c" [ 986.646468] env[63418]: _type = "Task" [ 986.646468] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.652733] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245406, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.658872] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5244102b-a988-5842-ced1-e0f574599f2c, 'name': SearchDatastore_Task, 'duration_secs': 0.009181} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.659171] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.659437] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.659679] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.659833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.660023] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.660269] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-004c943b-fd37-4f31-9afa-a0fc9a38af2a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.667501] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.667682] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.668414] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-183746b6-2f78-4fe7-b2a3-4307f86b2476 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.673534] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 986.673534] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e0e27d-12da-2d2e-96da-d341bc3230be" [ 986.673534] env[63418]: _type = "Task" [ 986.673534] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.680985] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e0e27d-12da-2d2e-96da-d341bc3230be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.816806] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245407, 'name': ReconfigVM_Task, 'duration_secs': 0.653581} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.817072] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 98f51fcd-951c-4c16-bc20-efc62ef359f5/98f51fcd-951c-4c16-bc20-efc62ef359f5.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.817695] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33a295ca-33c5-437d-9ea5-c5c85344ca88 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.824578] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 986.824578] env[63418]: value = "task-1245408" [ 986.824578] env[63418]: _type = "Task" [ 986.824578] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.835367] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245408, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.951583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.951964] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.951964] env[63418]: DEBUG nova.network.neutron [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.964871] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.965069] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.965366] env[63418]: DEBUG nova.network.neutron [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.965451] env[63418]: DEBUG nova.objects.instance [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'info_cache' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.142092] env[63418]: DEBUG nova.network.neutron [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [{"id": "492342f5-9181-4e37-9a04-ea9419e43523", "address": "fa:16:3e:6e:07:1a", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap492342f5-91", "ovs_interfaceid": "492342f5-9181-4e37-9a04-ea9419e43523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.147695] env[63418]: DEBUG nova.scheduler.client.report [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 987.156984] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245406, 'name': CloneVM_Task} progress is 95%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.184375] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e0e27d-12da-2d2e-96da-d341bc3230be, 'name': SearchDatastore_Task, 'duration_secs': 0.008658} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.185249] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2618ed7d-6f08-4e70-8610-321399be0af1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.190296] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 987.190296] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5215edbb-8825-9fee-c8e4-0f5d10c03e82" [ 987.190296] env[63418]: _type = "Task" [ 987.190296] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.198032] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5215edbb-8825-9fee-c8e4-0f5d10c03e82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.334187] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245408, 'name': Rename_Task, 'duration_secs': 0.173211} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.334570] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.334837] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19f70ad1-16d3-479c-865a-e8fb6314f2f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.341307] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 987.341307] env[63418]: value = "task-1245409" [ 987.341307] env[63418]: _type = "Task" [ 987.341307] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.350805] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245409, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.469207] env[63418]: DEBUG nova.objects.base [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 987.487256] env[63418]: DEBUG nova.network.neutron [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.647789] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.652450] env[63418]: DEBUG nova.network.neutron [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.653092] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.797s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.658420] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.617s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.659556] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245406, 'name': CloneVM_Task, 'duration_secs': 1.658093} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.659790] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Created linked-clone VM from snapshot [ 987.660543] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289e84f0-3590-4d67-8488-24dd4aed578f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.669497] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Uploading image e51a0692-2dc1-4746-a1c2-123e58c11a5c {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 987.676197] env[63418]: INFO nova.scheduler.client.report [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Deleted allocations for instance 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8 [ 987.685901] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 987.686249] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1a7e7cbb-fcdc-4ef3-8b40-09631fb09a90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.694786] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 987.694786] env[63418]: value = "task-1245410" [ 987.694786] env[63418]: _type = "Task" [ 987.694786] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.701341] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5215edbb-8825-9fee-c8e4-0f5d10c03e82, 'name': SearchDatastore_Task, 'duration_secs': 0.01041} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.701916] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.702216] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 3e645774-d1b3-468b-86a5-ff0c07e77f4e/3e645774-d1b3-468b-86a5-ff0c07e77f4e.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.702491] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f77d488-396b-4839-a604-aa9328a9991c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.707378] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245410, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.710919] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 987.710919] env[63418]: value = "task-1245411" [ 987.710919] env[63418]: _type = "Task" [ 987.710919] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.718581] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.851264] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245409, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.159870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.160164] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Instance network_info: |[{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 988.160586] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:9c:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6eb8c1f9-eec0-4763-8033-34a4efda65c1', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.168920] env[63418]: DEBUG oslo.service.loopingcall [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.172229] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.172723] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db5c835c-24e9-4de8-8f6b-d38e2a5ade3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.193272] env[63418]: DEBUG oslo_concurrency.lockutils [None req-42e496b4-3381-4dd4-81a6-6fcef7af1033 tempest-ServersNegativeTestJSON-2135634585 tempest-ServersNegativeTestJSON-2135634585-project-member] Lock "6465fb5c-7bc9-4197-b3fb-bad2c000a5f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.715s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.200650] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.200650] env[63418]: value = "task-1245412" [ 988.200650] env[63418]: _type = "Task" [ 988.200650] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.209521] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245410, 'name': Destroy_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.219533] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245412, 'name': CreateVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.227376] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467092} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.227679] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 3e645774-d1b3-468b-86a5-ff0c07e77f4e/3e645774-d1b3-468b-86a5-ff0c07e77f4e.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.227924] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.228255] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a88aa79a-0046-473d-a5bc-06caecb9c356 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.237735] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 988.237735] env[63418]: value = "task-1245413" [ 988.237735] env[63418]: _type = "Task" [ 988.237735] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.248761] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245413, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.344972] env[63418]: DEBUG nova.network.neutron [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [{"id": "01df43f5-4069-46a0-b1e6-4baff131e04e", "address": "fa:16:3e:69:39:54", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01df43f5-40", "ovs_interfaceid": "01df43f5-4069-46a0-b1e6-4baff131e04e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.356054] env[63418]: DEBUG oslo_vmware.api [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245409, 'name': PowerOnVM_Task, 'duration_secs': 0.642324} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.356333] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.356538] env[63418]: INFO nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Took 7.50 seconds to spawn the instance on the hypervisor. [ 988.356919] env[63418]: DEBUG nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 988.357929] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6e411e-973b-4a41-98cb-a90c7e7ddd42 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.402055] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e3eed8-9105-43cd-9166-e19730396ec8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.409692] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5040aa2-9925-4aca-b1b2-2a19770f3a22 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.444561] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da92d50e-d743-4e90-b1f9-5daefe509101 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.449677] env[63418]: DEBUG nova.compute.manager [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Received event network-changed-6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 988.449924] env[63418]: DEBUG nova.compute.manager [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Refreshing instance network info cache due to event network-changed-6eb8c1f9-eec0-4763-8033-34a4efda65c1. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 988.450558] env[63418]: DEBUG oslo_concurrency.lockutils [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.450558] env[63418]: DEBUG oslo_concurrency.lockutils [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.450558] env[63418]: DEBUG nova.network.neutron [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Refreshing network info cache for port 6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.457752] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63a2ce8-7544-47a5-a1ea-c45f3660f4c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.473066] env[63418]: DEBUG nova.compute.provider_tree [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.710479] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245410, 'name': Destroy_Task, 'duration_secs': 0.656638} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.714961] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Destroyed the VM [ 988.715319] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 988.715708] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245412, 'name': CreateVM_Task, 'duration_secs': 0.489656} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.715898] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-239962e3-c886-46ec-9344-3a64d31abef8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.717522] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.718607] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.718843] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.719209] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.719867] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80ee1980-345a-4694-9f5c-0ed1b94cd754 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.724618] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 988.724618] env[63418]: value = "task-1245414" [ 988.724618] env[63418]: _type = "Task" [ 988.724618] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.726613] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 988.726613] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528e3d43-6d56-8e62-ce51-17879c691031" [ 988.726613] env[63418]: _type = "Task" [ 988.726613] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.738683] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245414, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.745520] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528e3d43-6d56-8e62-ce51-17879c691031, 'name': SearchDatastore_Task, 'duration_secs': 0.010326} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.746033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.746413] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.746716] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.746950] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.747296] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.747970] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8fc6bd7-1278-4886-aa0f-4ce3ed8e0cd8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.752956] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245413, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072857} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.753534] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.754309] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f38e6a2-5bde-492f-bada-44e1d79774ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.758783] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.758966] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.768593] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6988c0a-956d-46cc-878c-e1d7887992de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.779824] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 3e645774-d1b3-468b-86a5-ff0c07e77f4e/3e645774-d1b3-468b-86a5-ff0c07e77f4e.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.780204] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdeb8586-ad75-4dbf-9969-ca1d2032c243 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.799464] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 988.799464] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c16d91-ee0a-a0ba-e1f1-6351a9f95f09" [ 988.799464] env[63418]: _type = "Task" [ 988.799464] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.800952] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 988.800952] env[63418]: value = "task-1245415" [ 988.800952] env[63418]: _type = "Task" [ 988.800952] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.813019] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c16d91-ee0a-a0ba-e1f1-6351a9f95f09, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.817807] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245415, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.818211] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c4b08f6-70b7-47cd-a409-afe3fab9dcd4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.824057] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 988.824057] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526f2e5d-7d3a-3aa0-62c2-509e9b193f98" [ 988.824057] env[63418]: _type = "Task" [ 988.824057] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.833013] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526f2e5d-7d3a-3aa0-62c2-509e9b193f98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.850854] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.883118] env[63418]: INFO nova.compute.manager [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Took 17.43 seconds to build instance. [ 988.976059] env[63418]: DEBUG nova.scheduler.client.report [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 989.166221] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab211964-3c76-487d-bb04-9005b3390b29 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.186491] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 0 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 989.240497] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245414, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.317532] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.337678] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526f2e5d-7d3a-3aa0-62c2-509e9b193f98, 'name': SearchDatastore_Task, 'duration_secs': 0.00905} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.338155] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.338546] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.338971] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6e7fc10-d256-4253-b21a-a2ef103bc68b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.346879] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 989.346879] env[63418]: value = "task-1245416" [ 989.346879] env[63418]: _type = "Task" [ 989.346879] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.357930] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245416, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.387236] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5fa789de-770f-4f9b-83bc-d7009ae62f80 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.943s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.695798] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.696414] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-424bf0ae-bb57-4692-bd5c-9e88edf8f590 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.705168] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 989.705168] env[63418]: value = "task-1245417" [ 989.705168] env[63418]: _type = "Task" [ 989.705168] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.724555] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.741669] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245414, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.815818] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245415, 'name': ReconfigVM_Task, 'duration_secs': 0.738258} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.816200] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 3e645774-d1b3-468b-86a5-ff0c07e77f4e/3e645774-d1b3-468b-86a5-ff0c07e77f4e.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.817162] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c704ff7b-b860-438b-a6e7-62836546f42c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.825278] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 989.825278] env[63418]: value = "task-1245418" [ 989.825278] env[63418]: _type = "Task" [ 989.825278] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.834462] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245418, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.842627] env[63418]: DEBUG nova.network.neutron [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updated VIF entry in instance network info cache for port 6eb8c1f9-eec0-4763-8033-34a4efda65c1. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.843044] env[63418]: DEBUG nova.network.neutron [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.857597] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245416, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475904} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.857854] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.858086] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.858354] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6374a9e0-07d4-42f4-bce8-7e277a6d2cde {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.861513] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.861804] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62c7b0be-7973-478d-ba82-6e314cfb157d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.865646] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 989.865646] env[63418]: value = "task-1245419" [ 989.865646] env[63418]: _type = "Task" [ 989.865646] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.869834] env[63418]: DEBUG oslo_vmware.api [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 989.869834] env[63418]: value = "task-1245420" [ 989.869834] env[63418]: _type = "Task" [ 989.869834] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.877086] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245419, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.882934] env[63418]: DEBUG oslo_vmware.api [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245420, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.989067] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.330s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.992570] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.984s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.993054] env[63418]: DEBUG nova.objects.instance [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'resources' on Instance uuid 0f99b32a-0125-4df0-919c-e5456b4ae4a5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.223248] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245417, 'name': PowerOffVM_Task, 'duration_secs': 0.277969} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.223556] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.223760] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 17 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 990.238277] env[63418]: DEBUG oslo_vmware.api [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245414, 'name': RemoveSnapshot_Task, 'duration_secs': 1.398937} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.240853] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 990.335292] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245418, 'name': Rename_Task, 'duration_secs': 0.250146} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.335605] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.335860] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27e1ff3d-b1b9-4820-b61a-cdec81c53d8a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.342352] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 990.342352] env[63418]: value = "task-1245421" [ 990.342352] env[63418]: _type = "Task" [ 990.342352] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.345603] env[63418]: DEBUG oslo_concurrency.lockutils [req-1e0d08ac-7541-49b5-bf35-3646ff8e90b1 req-91f11ba2-c10b-4a12-b6b9-34b8bc9bcbdd service nova] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.350401] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245421, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.376884] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245419, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079554} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.377563] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.378350] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73555312-fd29-4db4-b057-5dd99f047a35 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.383723] env[63418]: DEBUG oslo_vmware.api [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245420, 'name': PowerOnVM_Task, 'duration_secs': 0.448221} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.384565] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.384565] env[63418]: DEBUG nova.compute.manager [None req-c2909c7d-68fe-4707-b509-bc7f18da261d tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 990.385281] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cd46a7-c8b8-4b95-b388-4ad524fe6fa5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.406847] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.407587] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20d0444f-e02a-44e8-be13-052ff1f5ca47 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.435062] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 990.435062] env[63418]: value = "task-1245422" [ 990.435062] env[63418]: _type = "Task" [ 990.435062] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.443364] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.481815] env[63418]: DEBUG nova.compute.manager [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Received event network-changed-a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 990.482059] env[63418]: DEBUG nova.compute.manager [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Refreshing instance network info cache due to event network-changed-a6337c68-709c-4b53-b00a-b0e794cd45db. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 990.482262] env[63418]: DEBUG oslo_concurrency.lockutils [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] Acquiring lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.482404] env[63418]: DEBUG oslo_concurrency.lockutils [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] Acquired lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.482571] env[63418]: DEBUG nova.network.neutron [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Refreshing network info cache for port a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.578966] env[63418]: INFO nova.scheduler.client.report [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocation for migration 9856ef27-c1c8-4c00-8c0e-e947b31de4f1 [ 990.714052] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e899e13-b176-48a7-a3e1-916453e85405 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.723702] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123de571-1a47-413f-8226-2a2c2d05a7e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.735513] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.735889] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.736149] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.736452] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.736637] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.736813] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.737287] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.737287] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.737430] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.737537] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.737938] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.772285] env[63418]: WARNING nova.compute.manager [None req-3204d669-e4d2-42e8-b042-fae51e5585cd tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Image not found during snapshot: nova.exception.ImageNotFound: Image e51a0692-2dc1-4746-a1c2-123e58c11a5c could not be found. [ 990.773818] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbf38dd6-95be-4afb-aecf-5dde51eba5e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.787399] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6160c65-8516-46da-8fd5-fb885c429344 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.801137] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32001a62-5313-4cba-85ff-3c4b26ec374e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.809785] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 990.809785] env[63418]: value = "task-1245423" [ 990.809785] env[63418]: _type = "Task" [ 990.809785] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.820955] env[63418]: DEBUG nova.compute.provider_tree [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 990.828572] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245423, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.852417] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245421, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.947261] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.087355] env[63418]: DEBUG oslo_concurrency.lockutils [None req-0b54d1f6-70c0-4e73-bec7-0ff6f094512e tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.971s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.179902] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.180356] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.180717] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.181036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.181320] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.184434] env[63418]: INFO nova.compute.manager [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Terminating instance [ 991.315764] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245423, 'name': ReconfigVM_Task, 'duration_secs': 0.228718} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.316183] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 33 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.349315] env[63418]: ERROR nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [req-f39f7580-5475-4b0e-8e7d-fb8153160f9c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f39f7580-5475-4b0e-8e7d-fb8153160f9c"}]} [ 991.359012] env[63418]: DEBUG oslo_vmware.api [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245421, 'name': PowerOnVM_Task, 'duration_secs': 0.752917} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.359692] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.361133] env[63418]: INFO nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Took 8.19 seconds to spawn the instance on the hypervisor. [ 991.361133] env[63418]: DEBUG nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 991.361133] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e3f83f-2508-469d-b96a-4e7a50e17a2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.373060] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 991.393323] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 991.393567] env[63418]: DEBUG nova.compute.provider_tree [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 991.408791] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 991.430680] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 991.434933] env[63418]: DEBUG nova.network.neutron [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updated VIF entry in instance network info cache for port a6337c68-709c-4b53-b00a-b0e794cd45db. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.434933] env[63418]: DEBUG nova.network.neutron [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating instance_info_cache with network_info: [{"id": "a6337c68-709c-4b53-b00a-b0e794cd45db", "address": "fa:16:3e:78:ae:27", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6337c68-70", "ovs_interfaceid": "a6337c68-709c-4b53-b00a-b0e794cd45db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.450607] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245422, 'name': ReconfigVM_Task, 'duration_secs': 0.588246} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.451330] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.452215] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-712906df-a638-4fee-8134-703e45250517 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.459259] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 991.459259] env[63418]: value = "task-1245424" [ 991.459259] env[63418]: _type = "Task" [ 991.459259] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.471136] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245424, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.621743] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d24db4f-e680-4845-9a15-b0ed21ebb922 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.627158] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eb578e-264e-4eb8-a0a5-bd26617d8df1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.658151] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53f7908-a141-4e07-8044-302d604862cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.665101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3e07ea-5324-4562-9cf5-702b353f9909 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.678442] env[63418]: DEBUG nova.compute.provider_tree [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 991.691868] env[63418]: DEBUG nova.compute.manager [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 991.692085] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.692886] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3367f345-be7c-418c-ba1b-5b686a206d5c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.700212] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.700491] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ef19c0f-36fe-47ac-bc21-e4a0bb6b1ae3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.707021] env[63418]: DEBUG oslo_vmware.api [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 991.707021] env[63418]: value = "task-1245425" [ 991.707021] env[63418]: _type = "Task" [ 991.707021] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.714480] env[63418]: DEBUG oslo_vmware.api [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.727343] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0c016d4e-ca34-4831-a567-e794012681db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.727573] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0c016d4e-ca34-4831-a567-e794012681db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.727820] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "0c016d4e-ca34-4831-a567-e794012681db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.728046] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0c016d4e-ca34-4831-a567-e794012681db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.728225] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0c016d4e-ca34-4831-a567-e794012681db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.730379] env[63418]: INFO nova.compute.manager [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Terminating instance [ 991.824122] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.824437] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.824724] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.825061] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.825342] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.825620] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.826036] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.826351] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.826695] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.827059] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.827368] env[63418]: DEBUG nova.virt.hardware [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.832660] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 991.832972] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ebf5bda-e109-46ab-84f8-905c284afd35 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.851627] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 991.851627] env[63418]: value = "task-1245426" [ 991.851627] env[63418]: _type = "Task" [ 991.851627] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.860339] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245426, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.879253] env[63418]: INFO nova.compute.manager [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Took 16.56 seconds to build instance. [ 991.901559] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "0963bddc-6d3d-424d-a218-a2341fa9dab7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.901913] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.937398] env[63418]: DEBUG oslo_concurrency.lockutils [req-93b78c7e-3e30-4d93-8ac8-704302bb5651 req-9a081294-4a7f-4593-bcea-e3eb69a8a96d service nova] Releasing lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.972845] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245424, 'name': Rename_Task, 'duration_secs': 0.287233} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.973208] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 991.973468] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47f32a9e-4402-4cc8-95f4-b5fefa329154 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.980016] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 991.980016] env[63418]: value = "task-1245427" [ 991.980016] env[63418]: _type = "Task" [ 991.980016] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.987619] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245427, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.198640] env[63418]: ERROR nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [req-bfddd105-3af3-4348-9419-ca767cb74881] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bfddd105-3af3-4348-9419-ca767cb74881"}]} [ 992.216752] env[63418]: DEBUG oslo_vmware.api [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245425, 'name': PowerOffVM_Task, 'duration_secs': 0.283104} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.217663] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 992.219397] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.219590] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.220016] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a663bac-1468-4651-8b39-eeb110be155d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.230230] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 992.230432] env[63418]: DEBUG nova.compute.provider_tree [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 992.234687] env[63418]: DEBUG nova.compute.manager [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 992.234895] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.235698] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccd969b-f8f6-42c6-b3c1-1dd9b226a3a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.240059] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 992.245488] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.245703] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb25d1a4-81da-4eae-93bf-1de9e63f4849 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.250736] env[63418]: DEBUG oslo_vmware.api [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 992.250736] env[63418]: value = "task-1245429" [ 992.250736] env[63418]: _type = "Task" [ 992.250736] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.257890] env[63418]: DEBUG oslo_vmware.api [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.262668] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 992.296514] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.296777] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.296952] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore1] 5dac16e3-06a9-443a-90c8-9aacdd23fd91 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.297232] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd0f84d6-29a1-4301-9e43-fd761459ac0f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.303285] env[63418]: DEBUG oslo_vmware.api [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 992.303285] env[63418]: value = "task-1245430" [ 992.303285] env[63418]: _type = "Task" [ 992.303285] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.313707] env[63418]: DEBUG oslo_vmware.api [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245430, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.363478] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245426, 'name': ReconfigVM_Task, 'duration_secs': 0.167837} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.366163] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 992.367226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b007b31f-2b55-4227-b5aa-a62524159eaf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.387230] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5f54a37c-1260-4ddc-b27d-a6682e25ae2f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.073s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.396630] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6/680c94e8-c6d9-4aab-92d4-047a0e70eec6.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.400305] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af0205bf-dda0-4ba4-b280-578897b0e262 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.415287] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 992.423115] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 992.423115] env[63418]: value = "task-1245431" [ 992.423115] env[63418]: _type = "Task" [ 992.423115] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.434949] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245431, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.489972] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245427, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.491736] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e607d850-12e1-4e49-9527-4058ce70d7fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.499205] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada0b408-43e2-44d7-b012-d12a99d1ba52 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.532657] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12f412d-300d-4185-b8e6-728721e9eacf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.540947] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2415e8-795b-4000-b741-1d57931df2aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.555674] env[63418]: DEBUG nova.compute.provider_tree [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.761572] env[63418]: DEBUG oslo_vmware.api [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245429, 'name': PowerOffVM_Task, 'duration_secs': 0.176051} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.761572] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.761758] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.762019] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45ae6085-3aca-4613-86c7-71a8ae5c0704 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.813309] env[63418]: DEBUG oslo_vmware.api [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245430, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154699} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.813642] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.813850] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.814044] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.814231] env[63418]: INFO nova.compute.manager [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Took 1.12 seconds to destroy the instance on the hypervisor. [ 992.814477] env[63418]: DEBUG oslo.service.loopingcall [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.814743] env[63418]: DEBUG nova.compute.manager [-] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 992.814889] env[63418]: DEBUG nova.network.neutron [-] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.825675] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.825930] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.826128] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleting the datastore file [datastore2] 0c016d4e-ca34-4831-a567-e794012681db {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.826401] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29a37753-94b1-4999-b43b-e5fab93336a8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.833042] env[63418]: DEBUG oslo_vmware.api [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for the task: (returnval){ [ 992.833042] env[63418]: value = "task-1245433" [ 992.833042] env[63418]: _type = "Task" [ 992.833042] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.841896] env[63418]: DEBUG oslo_vmware.api [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245433, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.938365] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245431, 'name': ReconfigVM_Task, 'duration_secs': 0.27847} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.938824] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6/680c94e8-c6d9-4aab-92d4-047a0e70eec6.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.939182] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 50 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.949032] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.995056] env[63418]: DEBUG oslo_vmware.api [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245427, 'name': PowerOnVM_Task, 'duration_secs': 0.635049} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.995056] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.995056] env[63418]: INFO nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Took 7.46 seconds to spawn the instance on the hypervisor. [ 992.995056] env[63418]: DEBUG nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 992.995056] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b8b062-479d-406a-9ced-9faa076aeb6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.058785] env[63418]: DEBUG nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 993.187386] env[63418]: DEBUG nova.compute.manager [req-33f6da9e-d2f8-43db-9dc5-ffdfbdf6a74a req-9266553a-c5a5-4bde-a544-b21c4406239b service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Received event network-vif-deleted-7ab35b58-105e-4976-b7f7-05e0c77de3fa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 993.187556] env[63418]: INFO nova.compute.manager [req-33f6da9e-d2f8-43db-9dc5-ffdfbdf6a74a req-9266553a-c5a5-4bde-a544-b21c4406239b service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Neutron deleted interface 7ab35b58-105e-4976-b7f7-05e0c77de3fa; detaching it from the instance and deleting it from the info cache [ 993.187734] env[63418]: DEBUG nova.network.neutron [req-33f6da9e-d2f8-43db-9dc5-ffdfbdf6a74a req-9266553a-c5a5-4bde-a544-b21c4406239b service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.343675] env[63418]: DEBUG oslo_vmware.api [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Task: {'id': task-1245433, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.371315} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.343930] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.344134] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 993.344310] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 993.344552] env[63418]: INFO nova.compute.manager [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Took 1.11 seconds to destroy the instance on the hypervisor. [ 993.344795] env[63418]: DEBUG oslo.service.loopingcall [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.344996] env[63418]: DEBUG nova.compute.manager [-] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 993.345116] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c016d4e-ca34-4831-a567-e794012681db] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 993.445662] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9192991-a4bd-4a7a-8c59-80e7c6b62d2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.467515] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d3bac2-c539-4b0b-8d4f-25d77d9441f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.484863] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 67 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.509769] env[63418]: INFO nova.compute.manager [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Took 15.98 seconds to build instance. [ 993.569090] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.576s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.571282] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.624s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.572870] env[63418]: INFO nova.compute.claims [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.594257] env[63418]: INFO nova.scheduler.client.report [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted allocations for instance 0f99b32a-0125-4df0-919c-e5456b4ae4a5 [ 993.662977] env[63418]: DEBUG nova.network.neutron [-] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.690751] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8bdc7b2-7b78-43af-b4aa-68f787da8c14 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.700084] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc02fc3f-caa1-4551-9a59-5b5b4761e11a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.727764] env[63418]: DEBUG nova.compute.manager [req-33f6da9e-d2f8-43db-9dc5-ffdfbdf6a74a req-9266553a-c5a5-4bde-a544-b21c4406239b service nova] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Detach interface failed, port_id=7ab35b58-105e-4976-b7f7-05e0c77de3fa, reason: Instance 5dac16e3-06a9-443a-90c8-9aacdd23fd91 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 993.815915] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.816178] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.011770] env[63418]: DEBUG oslo_concurrency.lockutils [None req-13506949-3544-4384-8985-b6de201e2720 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.490s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.084516] env[63418]: DEBUG nova.network.neutron [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Port 492342f5-9181-4e37-9a04-ea9419e43523 binding to destination host cpu-1 is already ACTIVE {{(pid=63418) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 994.103692] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74751e72-7d41-4d03-897c-600796f757c1 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "0f99b32a-0125-4df0-919c-e5456b4ae4a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.121s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.113074] env[63418]: DEBUG nova.network.neutron [-] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.164979] env[63418]: INFO nova.compute.manager [-] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Took 1.35 seconds to deallocate network for instance. [ 994.316632] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.316933] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.317174] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.317365] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.317538] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.319319] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 994.322142] env[63418]: INFO nova.compute.manager [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Terminating instance [ 994.615684] env[63418]: INFO nova.compute.manager [-] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Took 1.27 seconds to deallocate network for instance. [ 994.670937] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.743285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05def374-abf5-45c1-823c-efaf6d495d3e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.751483] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050bf1a5-aaeb-4e83-a854-e842139dc003 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.782626] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67a70d8-502e-44e0-bfc2-d389be11f21a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.790486] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa01a63-2ecd-46fa-a772-0d02ffc8df2d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.805054] env[63418]: DEBUG nova.compute.provider_tree [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.828268] env[63418]: DEBUG nova.compute.manager [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 994.828429] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.829419] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248ff42c-bf9f-4c78-a184-99a0f5a92c5c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.836773] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.836993] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08f4cca4-7e8d-4e86-8f48-7950c5fd6185 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.838847] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.844153] env[63418]: DEBUG oslo_vmware.api [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 994.844153] env[63418]: value = "task-1245434" [ 994.844153] env[63418]: _type = "Task" [ 994.844153] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.851832] env[63418]: DEBUG oslo_vmware.api [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.109667] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.109667] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.109835] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.122581] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.215569] env[63418]: DEBUG nova.compute.manager [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Received event network-vif-deleted-050c0715-5f26-4662-9135-1f933411cf3e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 995.215817] env[63418]: DEBUG nova.compute.manager [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Received event network-changed-6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 995.215987] env[63418]: DEBUG nova.compute.manager [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Refreshing instance network info cache due to event network-changed-6eb8c1f9-eec0-4763-8033-34a4efda65c1. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 995.216248] env[63418]: DEBUG oslo_concurrency.lockutils [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.216400] env[63418]: DEBUG oslo_concurrency.lockutils [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.216581] env[63418]: DEBUG nova.network.neutron [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Refreshing network info cache for port 6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.308490] env[63418]: DEBUG nova.scheduler.client.report [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 995.354540] env[63418]: DEBUG oslo_vmware.api [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245434, 'name': PowerOffVM_Task, 'duration_secs': 0.347435} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.355105] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.355288] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.355567] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2f41b51-85fc-430b-a36d-d44ed563addc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.417093] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.417386] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.417611] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleting the datastore file [datastore1] 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.417930] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ca2cd35-9c68-4b55-aff8-b3f6e79cd8a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.424851] env[63418]: DEBUG oslo_vmware.api [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for the task: (returnval){ [ 995.424851] env[63418]: value = "task-1245436" [ 995.424851] env[63418]: _type = "Task" [ 995.424851] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.432472] env[63418]: DEBUG oslo_vmware.api [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245436, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.814059] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.814609] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 995.817300] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.146s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.817506] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.819685] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.981s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.821170] env[63418]: INFO nova.compute.claims [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.846555] env[63418]: INFO nova.scheduler.client.report [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocations for instance 5dac16e3-06a9-443a-90c8-9aacdd23fd91 [ 995.937709] env[63418]: DEBUG oslo_vmware.api [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Task: {'id': task-1245436, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.502648} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.937709] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.938036] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.938262] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.938498] env[63418]: INFO nova.compute.manager [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 995.939157] env[63418]: DEBUG oslo.service.loopingcall [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.939157] env[63418]: DEBUG nova.compute.manager [-] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 995.939157] env[63418]: DEBUG nova.network.neutron [-] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.966625] env[63418]: DEBUG nova.network.neutron [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updated VIF entry in instance network info cache for port 6eb8c1f9-eec0-4763-8033-34a4efda65c1. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.967089] env[63418]: DEBUG nova.network.neutron [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.148324] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.148549] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.148743] env[63418]: DEBUG nova.network.neutron [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.326077] env[63418]: DEBUG nova.compute.utils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.327582] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 996.327754] env[63418]: DEBUG nova.network.neutron [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.354753] env[63418]: DEBUG oslo_concurrency.lockutils [None req-54f495a7-8a62-4e09-90c2-fcd268a6a79b tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "5dac16e3-06a9-443a-90c8-9aacdd23fd91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.174s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.382873] env[63418]: DEBUG nova.compute.manager [req-e3cfb029-11eb-4269-b7f7-b50faf088626 req-4e7668e8-5ca8-4869-bc48-1a30455544fa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Received event network-vif-deleted-9b1f9284-9bb8-49b0-80f1-c2154e6ba534 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 996.383202] env[63418]: INFO nova.compute.manager [req-e3cfb029-11eb-4269-b7f7-b50faf088626 req-4e7668e8-5ca8-4869-bc48-1a30455544fa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Neutron deleted interface 9b1f9284-9bb8-49b0-80f1-c2154e6ba534; detaching it from the instance and deleting it from the info cache [ 996.383451] env[63418]: DEBUG nova.network.neutron [req-e3cfb029-11eb-4269-b7f7-b50faf088626 req-4e7668e8-5ca8-4869-bc48-1a30455544fa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.394633] env[63418]: DEBUG nova.policy [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ef8f62bd75a4d109a2e1881408754cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11fffb9badce4abeadce3ab70dff7d58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.469888] env[63418]: DEBUG oslo_concurrency.lockutils [req-55f554e3-6083-4df7-b6c4-6428484a04e7 req-3a85b279-f84c-43c8-b0bd-c5a375e3d2ef service nova] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.644118] env[63418]: DEBUG nova.network.neutron [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Successfully created port: a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.830895] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 996.846409] env[63418]: DEBUG nova.network.neutron [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [{"id": "492342f5-9181-4e37-9a04-ea9419e43523", "address": "fa:16:3e:6e:07:1a", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap492342f5-91", "ovs_interfaceid": "492342f5-9181-4e37-9a04-ea9419e43523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.859074] env[63418]: DEBUG nova.network.neutron [-] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.886487] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca56bee2-6715-4bca-ba50-4c98241046e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.898215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4373d622-6809-44e9-917e-554073d0e91e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.934456] env[63418]: DEBUG nova.compute.manager [req-e3cfb029-11eb-4269-b7f7-b50faf088626 req-4e7668e8-5ca8-4869-bc48-1a30455544fa service nova] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Detach interface failed, port_id=9b1f9284-9bb8-49b0-80f1-c2154e6ba534, reason: Instance 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 997.019101] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89915c5-a077-43ed-8ea3-20b96cd55134 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.026628] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66ac75f-d64b-48fd-a504-09a3b6cbf6d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.057911] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3f6551-5999-461b-9f95-36839c6263b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.065603] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1115f8e-fe7f-41e4-8713-a3326933958a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.078648] env[63418]: DEBUG nova.compute.provider_tree [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.350489] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.365209] env[63418]: INFO nova.compute.manager [-] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Took 1.43 seconds to deallocate network for instance. [ 997.581479] env[63418]: DEBUG nova.scheduler.client.report [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 997.845087] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 997.871979] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.872117] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.872219] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.872409] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.873017] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.874032] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.874032] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.874032] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.874032] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.874032] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.874314] env[63418]: DEBUG nova.virt.hardware [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.874977] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.875876] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b771bb7-c9d1-4a64-9705-018073eb59af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.879699] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b339bc-7e10-49b4-b0a4-c672b0ef3596 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.900683] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8b9c1b-c2ba-4a10-a175-16a18b439e80 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.905350] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5d061b-e6e0-4d19-8d35-b8278bb39ac0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.921703] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 83 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 998.027250] env[63418]: DEBUG nova.compute.manager [req-b11c60ac-d5ed-4cb9-a341-b37723dfe0b2 req-d58ece12-215f-4906-9af5-98b9a91a40e4 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Received event network-vif-plugged-a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 998.027479] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c60ac-d5ed-4cb9-a341-b37723dfe0b2 req-d58ece12-215f-4906-9af5-98b9a91a40e4 service nova] Acquiring lock "0963bddc-6d3d-424d-a218-a2341fa9dab7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.027697] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c60ac-d5ed-4cb9-a341-b37723dfe0b2 req-d58ece12-215f-4906-9af5-98b9a91a40e4 service nova] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.027910] env[63418]: DEBUG oslo_concurrency.lockutils [req-b11c60ac-d5ed-4cb9-a341-b37723dfe0b2 req-d58ece12-215f-4906-9af5-98b9a91a40e4 service nova] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.028099] env[63418]: DEBUG nova.compute.manager [req-b11c60ac-d5ed-4cb9-a341-b37723dfe0b2 req-d58ece12-215f-4906-9af5-98b9a91a40e4 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] No waiting events found dispatching network-vif-plugged-a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 998.028276] env[63418]: WARNING nova.compute.manager [req-b11c60ac-d5ed-4cb9-a341-b37723dfe0b2 req-d58ece12-215f-4906-9af5-98b9a91a40e4 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Received unexpected event network-vif-plugged-a268c9d0-c5bc-4d5d-8422-f95186e595e5 for instance with vm_state building and task_state spawning. [ 998.087440] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.088060] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 998.090622] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.968s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.090846] env[63418]: DEBUG nova.objects.instance [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lazy-loading 'resources' on Instance uuid 0c016d4e-ca34-4831-a567-e794012681db {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.173266] env[63418]: DEBUG nova.network.neutron [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Successfully updated port: a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.429298] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.429659] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-949a7678-6929-48cd-8cda-f75b7577250b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.437140] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 998.437140] env[63418]: value = "task-1245437" [ 998.437140] env[63418]: _type = "Task" [ 998.437140] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.446584] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.593569] env[63418]: DEBUG nova.compute.utils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.595455] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 998.595744] env[63418]: DEBUG nova.network.neutron [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.647243] env[63418]: DEBUG nova.policy [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.679314] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "refresh_cache-0963bddc-6d3d-424d-a218-a2341fa9dab7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.679500] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "refresh_cache-0963bddc-6d3d-424d-a218-a2341fa9dab7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.679663] env[63418]: DEBUG nova.network.neutron [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 998.760663] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3447677-5d51-4025-b9b0-3784b89cbe26 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.768973] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c49af2b-5ee7-4ffe-8b0f-7a09553566e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.802162] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c268bbc-7ad6-408f-98a7-4ef19968f475 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.810676] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416b7cca-8980-4f01-98fc-da8c366c886e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.828713] env[63418]: DEBUG nova.compute.provider_tree [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.947771] env[63418]: DEBUG oslo_vmware.api [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245437, 'name': PowerOnVM_Task, 'duration_secs': 0.478623} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.948264] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.948387] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-3382defd-867f-44d3-9125-7f184914f920 tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance '680c94e8-c6d9-4aab-92d4-047a0e70eec6' progress to 100 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 999.098864] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 999.124950] env[63418]: DEBUG nova.network.neutron [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Successfully created port: 24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 999.209980] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 999.210246] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 999.225225] env[63418]: DEBUG nova.network.neutron [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 999.331407] env[63418]: DEBUG nova.scheduler.client.report [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 999.360922] env[63418]: DEBUG nova.network.neutron [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Updating instance_info_cache with network_info: [{"id": "a268c9d0-c5bc-4d5d-8422-f95186e595e5", "address": "fa:16:3e:43:1e:f9", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa268c9d0-c5", "ovs_interfaceid": "a268c9d0-c5bc-4d5d-8422-f95186e595e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.839798] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.842241] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.967s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.842484] env[63418]: DEBUG nova.objects.instance [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lazy-loading 'resources' on Instance uuid 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.862475] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "refresh_cache-0963bddc-6d3d-424d-a218-a2341fa9dab7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.862800] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Instance network_info: |[{"id": "a268c9d0-c5bc-4d5d-8422-f95186e595e5", "address": "fa:16:3e:43:1e:f9", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa268c9d0-c5", "ovs_interfaceid": "a268c9d0-c5bc-4d5d-8422-f95186e595e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 999.863244] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:1e:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a268c9d0-c5bc-4d5d-8422-f95186e595e5', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.871113] env[63418]: DEBUG oslo.service.loopingcall [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.872123] env[63418]: INFO nova.scheduler.client.report [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Deleted allocations for instance 0c016d4e-ca34-4831-a567-e794012681db [ 999.873248] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.875563] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5d3413c-900b-4faa-99ef-1d71c38ec499 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.895827] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.895827] env[63418]: value = "task-1245438" [ 999.895827] env[63418]: _type = "Task" [ 999.895827] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.903316] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245438, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.054806] env[63418]: DEBUG nova.compute.manager [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Received event network-changed-a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1000.055046] env[63418]: DEBUG nova.compute.manager [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Refreshing instance network info cache due to event network-changed-a268c9d0-c5bc-4d5d-8422-f95186e595e5. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1000.055208] env[63418]: DEBUG oslo_concurrency.lockutils [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] Acquiring lock "refresh_cache-0963bddc-6d3d-424d-a218-a2341fa9dab7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.055359] env[63418]: DEBUG oslo_concurrency.lockutils [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] Acquired lock "refresh_cache-0963bddc-6d3d-424d-a218-a2341fa9dab7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.055520] env[63418]: DEBUG nova.network.neutron [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Refreshing network info cache for port a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.113906] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1000.142314] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.142678] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.142855] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.143055] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.143209] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.143358] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.143570] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.143730] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.143897] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.144125] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.144323] env[63418]: DEBUG nova.virt.hardware [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.145246] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282e8414-cc2b-47cd-bd2b-5c501663d929 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.153150] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb903eb-cf39-4dc9-94a7-b97a2847b976 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.394197] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f2872332-7401-4bb6-aceb-05b1f052a078 tempest-ImagesTestJSON-1758946497 tempest-ImagesTestJSON-1758946497-project-member] Lock "0c016d4e-ca34-4831-a567-e794012681db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.666s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.406343] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245438, 'name': CreateVM_Task, 'duration_secs': 0.322613} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.409214] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1000.409971] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.410150] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.410946] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.410946] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c728edf-8e13-4054-bfe6-11662b2e44d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.416902] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1000.416902] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52accec8-aee0-acc7-28b0-213dee010c04" [ 1000.416902] env[63418]: _type = "Task" [ 1000.416902] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.427938] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52accec8-aee0-acc7-28b0-213dee010c04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.512926] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeaf620c-7450-4108-859d-389c5b425935 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.521471] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ceb734-943e-4f62-89ac-8420ca041009 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.551018] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8c759b-6c17-4b19-8213-7508bd7f0494 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.560945] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f3dd46-8d34-4088-8ac6-3eb51ab76c2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.576436] env[63418]: DEBUG nova.compute.provider_tree [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.686743] env[63418]: DEBUG nova.network.neutron [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Successfully updated port: 24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.909876] env[63418]: DEBUG nova.network.neutron [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Updated VIF entry in instance network info cache for port a268c9d0-c5bc-4d5d-8422-f95186e595e5. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.910291] env[63418]: DEBUG nova.network.neutron [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Updating instance_info_cache with network_info: [{"id": "a268c9d0-c5bc-4d5d-8422-f95186e595e5", "address": "fa:16:3e:43:1e:f9", "network": {"id": "bc7edc6c-2c09-4245-ab05-ba3fe005e969", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2038786037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11fffb9badce4abeadce3ab70dff7d58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa268c9d0-c5", "ovs_interfaceid": "a268c9d0-c5bc-4d5d-8422-f95186e595e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.928582] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52accec8-aee0-acc7-28b0-213dee010c04, 'name': SearchDatastore_Task, 'duration_secs': 0.010137} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.928817] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.929065] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.929310] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.929463] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.929648] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.929916] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af28cc03-22e0-47be-a1f3-752af8409f79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.938613] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.938834] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.939658] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab4a8382-d217-4257-8ee3-1f5657986dac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.945729] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1000.945729] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a79ef1-8a71-9a68-5f5f-e9622022513e" [ 1000.945729] env[63418]: _type = "Task" [ 1000.945729] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.954039] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a79ef1-8a71-9a68-5f5f-e9622022513e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.079556] env[63418]: DEBUG nova.scheduler.client.report [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1001.189715] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-9ae2cb1e-3999-4471-8a81-7f86db857ff3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.189843] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-9ae2cb1e-3999-4471-8a81-7f86db857ff3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.190054] env[63418]: DEBUG nova.network.neutron [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.209875] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1001.210106] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 1001.345027] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.345027] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.345027] env[63418]: DEBUG nova.compute.manager [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Going to confirm migration 4 {{(pid=63418) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 1001.416056] env[63418]: DEBUG oslo_concurrency.lockutils [req-1cb32701-72ee-4c71-97f6-4d8aac45f3fe req-ba095974-0ac8-4963-8fc3-aac9b33308a1 service nova] Releasing lock "refresh_cache-0963bddc-6d3d-424d-a218-a2341fa9dab7" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.457918] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a79ef1-8a71-9a68-5f5f-e9622022513e, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.461046] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f865098c-9f8e-4726-bf00-587e7cc3d519 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.464269] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1001.464269] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523eaace-de8d-2c00-d7b9-a4d8bb32f7f8" [ 1001.464269] env[63418]: _type = "Task" [ 1001.464269] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.471958] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523eaace-de8d-2c00-d7b9-a4d8bb32f7f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.587298] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.606200] env[63418]: INFO nova.scheduler.client.report [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Deleted allocations for instance 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5 [ 1001.732876] env[63418]: DEBUG nova.network.neutron [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.898338] env[63418]: DEBUG nova.network.neutron [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Updating instance_info_cache with network_info: [{"id": "24cd6ab7-ed75-4c98-9839-27d80bd00464", "address": "fa:16:3e:4c:b4:6c", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24cd6ab7-ed", "ovs_interfaceid": "24cd6ab7-ed75-4c98-9839-27d80bd00464", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.936649] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.939462] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquired lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.939462] env[63418]: DEBUG nova.network.neutron [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.939462] env[63418]: DEBUG nova.objects.instance [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'info_cache' on Instance uuid 680c94e8-c6d9-4aab-92d4-047a0e70eec6 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.978687] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523eaace-de8d-2c00-d7b9-a4d8bb32f7f8, 'name': SearchDatastore_Task, 'duration_secs': 0.009349} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.980072] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.980072] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 0963bddc-6d3d-424d-a218-a2341fa9dab7/0963bddc-6d3d-424d-a218-a2341fa9dab7.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1001.980479] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-303b4205-7b72-493c-af11-5c940d4015e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.989138] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1001.989138] env[63418]: value = "task-1245439" [ 1001.989138] env[63418]: _type = "Task" [ 1001.989138] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.997463] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.104246] env[63418]: DEBUG nova.compute.manager [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Received event network-vif-plugged-24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1002.104246] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] Acquiring lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.104246] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.104246] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.104894] env[63418]: DEBUG nova.compute.manager [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] No waiting events found dispatching network-vif-plugged-24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1002.105167] env[63418]: WARNING nova.compute.manager [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Received unexpected event network-vif-plugged-24cd6ab7-ed75-4c98-9839-27d80bd00464 for instance with vm_state building and task_state spawning. [ 1002.105350] env[63418]: DEBUG nova.compute.manager [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Received event network-changed-24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1002.105525] env[63418]: DEBUG nova.compute.manager [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Refreshing instance network info cache due to event network-changed-24cd6ab7-ed75-4c98-9839-27d80bd00464. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1002.105736] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] Acquiring lock "refresh_cache-9ae2cb1e-3999-4471-8a81-7f86db857ff3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.114594] env[63418]: DEBUG oslo_concurrency.lockutils [None req-312a0d58-40c4-4c43-91c8-ac9b3cd441f6 tempest-AttachInterfacesTestJSON-1462268461 tempest-AttachInterfacesTestJSON-1462268461-project-member] Lock "039f4cf6-137b-4356-a1b8-dd47f8f0d6f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.798s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.403860] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-9ae2cb1e-3999-4471-8a81-7f86db857ff3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.404471] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Instance network_info: |[{"id": "24cd6ab7-ed75-4c98-9839-27d80bd00464", "address": "fa:16:3e:4c:b4:6c", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24cd6ab7-ed", "ovs_interfaceid": "24cd6ab7-ed75-4c98-9839-27d80bd00464", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1002.404919] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] Acquired lock "refresh_cache-9ae2cb1e-3999-4471-8a81-7f86db857ff3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.405137] env[63418]: DEBUG nova.network.neutron [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Refreshing network info cache for port 24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.407061] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:b4:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24cd6ab7-ed75-4c98-9839-27d80bd00464', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.418394] env[63418]: DEBUG oslo.service.loopingcall [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.418943] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.419216] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73a392e1-046a-4342-941f-fe0a9d82c39b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.439886] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.439886] env[63418]: value = "task-1245440" [ 1002.439886] env[63418]: _type = "Task" [ 1002.439886] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.448283] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245440, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.498085] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245439, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457821} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.498350] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 0963bddc-6d3d-424d-a218-a2341fa9dab7/0963bddc-6d3d-424d-a218-a2341fa9dab7.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.498565] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.498818] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7a8fb87-b673-498d-b2fe-37b831a56f00 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.504702] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1002.504702] env[63418]: value = "task-1245441" [ 1002.504702] env[63418]: _type = "Task" [ 1002.504702] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.512135] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245441, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.953224] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245440, 'name': CreateVM_Task, 'duration_secs': 0.466447} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.953397] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.954297] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.954479] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.954828] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.955094] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-000daa19-2966-4da0-a511-71ba04a5d68c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.959259] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1002.959259] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526650ad-3c6a-7c4c-4c35-7581a8d66d9f" [ 1002.959259] env[63418]: _type = "Task" [ 1002.959259] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.966481] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526650ad-3c6a-7c4c-4c35-7581a8d66d9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.013096] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245441, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067058} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.013363] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.014120] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014536ae-60b1-4838-80c9-d0fd26615c6f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.039098] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 0963bddc-6d3d-424d-a218-a2341fa9dab7/0963bddc-6d3d-424d-a218-a2341fa9dab7.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.042710] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3277c68d-3f13-4ec4-aef6-6d183ad46c2d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.063099] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1003.063099] env[63418]: value = "task-1245442" [ 1003.063099] env[63418]: _type = "Task" [ 1003.063099] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.069681] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245442, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.181750] env[63418]: DEBUG nova.network.neutron [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [{"id": "492342f5-9181-4e37-9a04-ea9419e43523", "address": "fa:16:3e:6e:07:1a", "network": {"id": "7b98992f-6a4f-4002-af3d-aa5630c6e706", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1482441249-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6d52f1fceb24234a8d967038b43c857", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap492342f5-91", "ovs_interfaceid": "492342f5-9181-4e37-9a04-ea9419e43523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.306190] env[63418]: DEBUG nova.network.neutron [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Updated VIF entry in instance network info cache for port 24cd6ab7-ed75-4c98-9839-27d80bd00464. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.306679] env[63418]: DEBUG nova.network.neutron [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Updating instance_info_cache with network_info: [{"id": "24cd6ab7-ed75-4c98-9839-27d80bd00464", "address": "fa:16:3e:4c:b4:6c", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24cd6ab7-ed", "ovs_interfaceid": "24cd6ab7-ed75-4c98-9839-27d80bd00464", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.469362] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526650ad-3c6a-7c4c-4c35-7581a8d66d9f, 'name': SearchDatastore_Task, 'duration_secs': 0.009145} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.469676] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.469917] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.470175] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.470326] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.470517] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.470777] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffd98c11-ac59-40d0-b51d-1f7198e66c8f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.478473] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.478647] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.479614] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7978bec6-d15c-465c-9fd5-e382e66f821f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.483957] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1003.483957] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527fb037-bf75-7911-a67f-85978fdd6d56" [ 1003.483957] env[63418]: _type = "Task" [ 1003.483957] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.491331] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527fb037-bf75-7911-a67f-85978fdd6d56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.571623] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245442, 'name': ReconfigVM_Task, 'duration_secs': 0.282584} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.571929] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 0963bddc-6d3d-424d-a218-a2341fa9dab7/0963bddc-6d3d-424d-a218-a2341fa9dab7.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.572647] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-105da01d-6eb4-493f-a91d-eade377224f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.579513] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1003.579513] env[63418]: value = "task-1245443" [ 1003.579513] env[63418]: _type = "Task" [ 1003.579513] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.588191] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245443, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.685150] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Releasing lock "refresh_cache-680c94e8-c6d9-4aab-92d4-047a0e70eec6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.685428] env[63418]: DEBUG nova.objects.instance [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'migration_context' on Instance uuid 680c94e8-c6d9-4aab-92d4-047a0e70eec6 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.729256] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Didn't find any instances for network info cache update. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10364}} [ 1003.729831] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.730030] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.730171] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 1003.730319] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.809306] env[63418]: DEBUG oslo_concurrency.lockutils [req-f7fb3596-b491-4f9b-8dc2-238799a9b823 req-9dea4a15-7a08-4af6-923d-f79a14577d09 service nova] Releasing lock "refresh_cache-9ae2cb1e-3999-4471-8a81-7f86db857ff3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.995967] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527fb037-bf75-7911-a67f-85978fdd6d56, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.996751] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b7217a8-89bb-4807-b58a-4435113858b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.001642] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1004.001642] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c6041c-4c31-1a7a-1faf-936c9bd6c6bd" [ 1004.001642] env[63418]: _type = "Task" [ 1004.001642] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.010043] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c6041c-4c31-1a7a-1faf-936c9bd6c6bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.089359] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245443, 'name': Rename_Task, 'duration_secs': 0.152325} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.089885] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1004.090241] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-649e6a5e-80b3-4297-8444-ddabb8614f38 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.099051] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1004.099051] env[63418]: value = "task-1245444" [ 1004.099051] env[63418]: _type = "Task" [ 1004.099051] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.104318] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.188858] env[63418]: DEBUG nova.objects.base [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Object Instance<680c94e8-c6d9-4aab-92d4-047a0e70eec6> lazy-loaded attributes: info_cache,migration_context {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1004.191292] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cd50b8-4cae-4391-93d9-c23688207fe1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.211441] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-511a9582-4851-431f-879d-c89caf8d1d43 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.220017] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 1004.220017] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d4d8a6-59d4-88be-169f-09e2a50204c8" [ 1004.220017] env[63418]: _type = "Task" [ 1004.220017] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.226965] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d4d8a6-59d4-88be-169f-09e2a50204c8, 'name': SearchDatastore_Task, 'duration_secs': 0.006606} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.227439] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.227788] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.234080] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.374564] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.377268] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.377268] env[63418]: INFO nova.compute.manager [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Shelving [ 1004.511940] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c6041c-4c31-1a7a-1faf-936c9bd6c6bd, 'name': SearchDatastore_Task, 'duration_secs': 0.009553} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.512041] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.512250] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 9ae2cb1e-3999-4471-8a81-7f86db857ff3/9ae2cb1e-3999-4471-8a81-7f86db857ff3.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.512532] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a35673a-c96b-4225-9601-cb26de455370 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.519179] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1004.519179] env[63418]: value = "task-1245445" [ 1004.519179] env[63418]: _type = "Task" [ 1004.519179] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.526732] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.606740] env[63418]: DEBUG oslo_vmware.api [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245444, 'name': PowerOnVM_Task, 'duration_secs': 0.483282} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.607088] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.607328] env[63418]: INFO nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Took 6.76 seconds to spawn the instance on the hypervisor. [ 1004.607534] env[63418]: DEBUG nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1004.608446] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c39b7c-7ce9-4306-8dbb-9be23e987201 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.888337] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cef06c-4357-4762-b4d3-1dbabb6e528c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.897639] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77d3d62-d0bb-421d-b758-13e2aa843b02 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.931358] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ce8c22-9f5c-447c-9fe1-34be3a9a8676 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.940533] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d99d05a-9a5b-497c-8740-048ff356b65c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.955305] env[63418]: DEBUG nova.compute.provider_tree [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.029577] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245445, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.127154] env[63418]: INFO nova.compute.manager [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Took 12.21 seconds to build instance. [ 1005.395342] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.395342] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bbe9330-bf21-4b38-bf6f-cf4e0a967cf2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.402153] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1005.402153] env[63418]: value = "task-1245446" [ 1005.402153] env[63418]: _type = "Task" [ 1005.402153] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.411270] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.478943] env[63418]: ERROR nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [req-e7331055-ab55-4e74-a2d0-6bd36752d532] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e7331055-ab55-4e74-a2d0-6bd36752d532"}]} [ 1005.497542] env[63418]: DEBUG nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1005.512549] env[63418]: DEBUG nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1005.512697] env[63418]: DEBUG nova.compute.provider_tree [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.525289] env[63418]: DEBUG nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1005.531045] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.696663} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.531350] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 9ae2cb1e-3999-4471-8a81-7f86db857ff3/9ae2cb1e-3999-4471-8a81-7f86db857ff3.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.531641] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.531988] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2ec0329-bd6b-4409-a114-61718a6656cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.539822] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1005.539822] env[63418]: value = "task-1245447" [ 1005.539822] env[63418]: _type = "Task" [ 1005.539822] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.549496] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245447, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.552732] env[63418]: DEBUG nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1005.629416] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b891eade-dd90-49cf-9dfb-5cef05e76480 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.727s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.702730] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b14563-b18c-4c30-b7c7-9809aff5624c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.711881] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93648770-673e-4ed8-98dd-6833b9e6bcdb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.744462] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fe05c1-0e9b-4485-ac5c-3a6d8d1f1c56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.752725] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a355d356-9081-467f-84b7-45402e53d9ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.767418] env[63418]: DEBUG nova.compute.provider_tree [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.914266] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245446, 'name': PowerOffVM_Task, 'duration_secs': 0.239021} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.914561] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.915447] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2add213-50c8-40da-af7e-30b40e2cdeb9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.934370] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5279db-2acb-4514-b142-d3f737302b20 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.051100] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245447, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07079} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.051380] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.052222] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924287fb-6a2d-49c5-a9eb-222d42a5ddf4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.076737] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 9ae2cb1e-3999-4471-8a81-7f86db857ff3/9ae2cb1e-3999-4471-8a81-7f86db857ff3.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.077114] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-833fb36a-e18f-437e-81a9-ad9185ced43a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.098570] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1006.098570] env[63418]: value = "task-1245448" [ 1006.098570] env[63418]: _type = "Task" [ 1006.098570] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.107399] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245448, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.307062] env[63418]: DEBUG nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 128 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1006.307062] env[63418]: DEBUG nova.compute.provider_tree [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 128 to 129 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1006.307062] env[63418]: DEBUG nova.compute.provider_tree [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.452030] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1006.452030] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bbef84d1-68b6-4b1b-aa19-938d7b1b9f83 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.460245] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1006.460245] env[63418]: value = "task-1245449" [ 1006.460245] env[63418]: _type = "Task" [ 1006.460245] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.475675] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245449, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.611026] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245448, 'name': ReconfigVM_Task, 'duration_secs': 0.403386} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.611026] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 9ae2cb1e-3999-4471-8a81-7f86db857ff3/9ae2cb1e-3999-4471-8a81-7f86db857ff3.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.613494] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c39e6612-1912-4d4f-91ec-4dc6636c4e9b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.618880] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1006.618880] env[63418]: value = "task-1245450" [ 1006.618880] env[63418]: _type = "Task" [ 1006.618880] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.629392] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245450, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.749987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "0963bddc-6d3d-424d-a218-a2341fa9dab7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.749987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.749987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "0963bddc-6d3d-424d-a218-a2341fa9dab7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.749987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.749987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.751137] env[63418]: INFO nova.compute.manager [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Terminating instance [ 1006.971974] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245449, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.131650] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245450, 'name': Rename_Task, 'duration_secs': 0.159978} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.132072] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.132419] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-558d55e3-81b7-4b77-a39c-287ad102bea2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.141251] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1007.141251] env[63418]: value = "task-1245451" [ 1007.141251] env[63418]: _type = "Task" [ 1007.141251] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.153916] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.260913] env[63418]: DEBUG nova.compute.manager [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1007.261399] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.263478] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573e4464-5285-4e0b-ae84-74088bda24f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.274243] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.274567] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26482fd1-50b5-444f-a725-4a219277c0b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.283687] env[63418]: DEBUG oslo_vmware.api [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1007.283687] env[63418]: value = "task-1245452" [ 1007.283687] env[63418]: _type = "Task" [ 1007.283687] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.298127] env[63418]: DEBUG oslo_vmware.api [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.320230] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.092s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.327656] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.093s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.328865] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.328865] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1007.329218] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270dd67d-c145-4800-99fa-840a46f95c88 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.338656] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9e2e61-b1ad-41ff-89c1-7f2081944faf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.354214] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33a23bb-4dda-4e84-8fb7-65d57aaf5d62 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.362391] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55ff4d7-4e6d-440f-9223-83cf314ebbf0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.394079] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180481MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1007.394187] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.394374] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.471976] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245449, 'name': CreateSnapshot_Task, 'duration_secs': 0.824369} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.473204] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1007.473314] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7529da-3c50-4a29-9993-eb5eb7b2b680 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.653088] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245451, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.793687] env[63418]: DEBUG oslo_vmware.api [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245452, 'name': PowerOffVM_Task, 'duration_secs': 0.360243} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.794013] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.794195] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.794455] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3322720-a26f-440e-af28-a42149d09027 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.872501] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.872656] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.872712] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleting the datastore file [datastore1] 0963bddc-6d3d-424d-a218-a2341fa9dab7 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.873070] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7755b1b3-9bd1-40a5-bff3-21099743045b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.882458] env[63418]: DEBUG oslo_vmware.api [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for the task: (returnval){ [ 1007.882458] env[63418]: value = "task-1245454" [ 1007.882458] env[63418]: _type = "Task" [ 1007.882458] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.897638] env[63418]: DEBUG oslo_vmware.api [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.920704] env[63418]: INFO nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocation for migration 1a492728-e2b4-48fb-bec5-473405172064 [ 1007.996377] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1007.997087] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9efe1122-1857-4a80-bbd1-c1f2da5e3d3a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.009624] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1008.009624] env[63418]: value = "task-1245455" [ 1008.009624] env[63418]: _type = "Task" [ 1008.009624] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.019517] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245455, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.153848] env[63418]: DEBUG oslo_vmware.api [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245451, 'name': PowerOnVM_Task, 'duration_secs': 0.628899} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.154200] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1008.154520] env[63418]: INFO nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Took 8.04 seconds to spawn the instance on the hypervisor. [ 1008.154749] env[63418]: DEBUG nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1008.155599] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb26bc49-1f65-48ad-be3d-0e52a06d916d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.393161] env[63418]: DEBUG oslo_vmware.api [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Task: {'id': task-1245454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335987} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.393319] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.393423] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.393718] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.393806] env[63418]: INFO nova.compute.manager [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1008.394045] env[63418]: DEBUG oslo.service.loopingcall [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.394272] env[63418]: DEBUG nova.compute.manager [-] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1008.394979] env[63418]: DEBUG nova.network.neutron [-] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.423596] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance bf091892-fefb-49dd-9416-708a06b35798 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.423744] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance cbcbaf4f-ee24-4072-83ae-ffde59478928 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.423921] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance d52db46b-2461-4bd0-be57-d414250aac7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.424097] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 98f51fcd-951c-4c16-bc20-efc62ef359f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.424221] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 3e645774-d1b3-468b-86a5-ff0c07e77f4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.424336] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 5a8b347c-e9ec-4355-bcf7-b7d205da194c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.424449] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 680c94e8-c6d9-4aab-92d4-047a0e70eec6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.424592] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 0963bddc-6d3d-424d-a218-a2341fa9dab7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.424900] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 9ae2cb1e-3999-4471-8a81-7f86db857ff3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.425189] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1008.425345] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1008.428689] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.085s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.523584] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245455, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.565998] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8350392c-7a02-459e-8ab3-5215ab63bf51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.574387] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f42912-351f-4953-a3ec-d4c52e5b7eae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.608020] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642a0e68-fa30-4338-aa95-a35f69e4c2fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.617064] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b36b0b-3d71-4f25-9647-00f63eacc4c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.631179] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.674214] env[63418]: INFO nova.compute.manager [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Took 13.85 seconds to build instance. [ 1009.022416] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245455, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.049400] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "80777e38-728b-4d2d-9003-f55ea7999299" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.049678] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "80777e38-728b-4d2d-9003-f55ea7999299" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.086251] env[63418]: DEBUG nova.compute.manager [req-4629b037-7428-45a0-80f0-382b40aa9677 req-a5431fbb-c805-48bf-8895-4baa6aac300a service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Received event network-vif-deleted-a268c9d0-c5bc-4d5d-8422-f95186e595e5 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1009.086461] env[63418]: INFO nova.compute.manager [req-4629b037-7428-45a0-80f0-382b40aa9677 req-a5431fbb-c805-48bf-8895-4baa6aac300a service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Neutron deleted interface a268c9d0-c5bc-4d5d-8422-f95186e595e5; detaching it from the instance and deleting it from the info cache [ 1009.086737] env[63418]: DEBUG nova.network.neutron [req-4629b037-7428-45a0-80f0-382b40aa9677 req-a5431fbb-c805-48bf-8895-4baa6aac300a service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.133839] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1009.175186] env[63418]: DEBUG oslo_concurrency.lockutils [None req-27ce2e63-86c0-4859-9710-d12610b75f0f tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.359s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.425520] env[63418]: DEBUG nova.network.neutron [-] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.531174] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245455, 'name': CloneVM_Task, 'duration_secs': 1.458206} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.532369] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Created linked-clone VM from snapshot [ 1009.534966] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e531f47-b539-4a84-8a86-388764643fa8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.551889] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Uploading image 8691809c-e40f-490b-bcba-aaf810e4f073 {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1009.555685] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1009.591211] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e28d520e-19b3-425b-9317-50ba23ce2ec3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.594367] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1009.594367] env[63418]: value = "vm-268530" [ 1009.594367] env[63418]: _type = "VirtualMachine" [ 1009.594367] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1009.594970] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9f937584-7db9-4b10-a800-ea0249cd9b0a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.606454] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afeef546-a42d-415b-bab2-1b1c797590aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.620984] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease: (returnval){ [ 1009.620984] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523f1f07-10a7-dc20-b404-4e30bb953db5" [ 1009.620984] env[63418]: _type = "HttpNfcLease" [ 1009.620984] env[63418]: } obtained for exporting VM: (result){ [ 1009.620984] env[63418]: value = "vm-268530" [ 1009.620984] env[63418]: _type = "VirtualMachine" [ 1009.620984] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1009.623269] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the lease: (returnval){ [ 1009.623269] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523f1f07-10a7-dc20-b404-4e30bb953db5" [ 1009.623269] env[63418]: _type = "HttpNfcLease" [ 1009.623269] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1009.630774] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1009.630774] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523f1f07-10a7-dc20-b404-4e30bb953db5" [ 1009.630774] env[63418]: _type = "HttpNfcLease" [ 1009.630774] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1009.631378] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1009.631378] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523f1f07-10a7-dc20-b404-4e30bb953db5" [ 1009.631378] env[63418]: _type = "HttpNfcLease" [ 1009.631378] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1009.632415] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edb5b50-de96-4c3e-90fd-3424f55f5a7a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.644311] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1009.644767] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.250s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.645263] env[63418]: DEBUG nova.compute.manager [req-4629b037-7428-45a0-80f0-382b40aa9677 req-a5431fbb-c805-48bf-8895-4baa6aac300a service nova] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Detach interface failed, port_id=a268c9d0-c5bc-4d5d-8422-f95186e595e5, reason: Instance 0963bddc-6d3d-424d-a218-a2341fa9dab7 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1009.652105] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525832df-8534-18db-d8d2-53aee909f2f6/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1009.652482] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525832df-8534-18db-d8d2-53aee909f2f6/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1009.744487] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-74326731-6b49-4e23-a255-48ab9313fe16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.886841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.886841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.886841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.886841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.886841] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.887681] env[63418]: INFO nova.compute.manager [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Terminating instance [ 1009.928436] env[63418]: INFO nova.compute.manager [-] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Took 1.53 seconds to deallocate network for instance. [ 1010.021986] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.022246] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.022523] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.022860] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.023071] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.026162] env[63418]: INFO nova.compute.manager [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Terminating instance [ 1010.076839] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.077252] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.078840] env[63418]: INFO nova.compute.claims [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.125678] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1010.125936] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1010.126267] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1010.394661] env[63418]: DEBUG nova.compute.manager [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1010.395364] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.396292] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25de5e33-7a3b-4523-a042-e5220b373190 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.406979] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.407730] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fb3254b-f6f8-4ef8-96c2-2037b95097c6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.415910] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 1010.415910] env[63418]: value = "task-1245457" [ 1010.415910] env[63418]: _type = "Task" [ 1010.415910] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.425912] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.434334] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.530780] env[63418]: DEBUG nova.compute.manager [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1010.531335] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.532639] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a02aef-9f37-44eb-a2f6-b099dda2f606 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.543147] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.543258] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e5de86f-5889-4b96-b0a5-bd9efa684733 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.553731] env[63418]: DEBUG oslo_vmware.api [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1010.553731] env[63418]: value = "task-1245458" [ 1010.553731] env[63418]: _type = "Task" [ 1010.553731] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.564170] env[63418]: DEBUG oslo_vmware.api [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245458, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.926924] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245457, 'name': PowerOffVM_Task, 'duration_secs': 0.350155} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.927292] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.927547] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.927955] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dbd2f3f-735f-4228-90d1-1726244e48cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.008247] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.008438] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.008632] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleting the datastore file [datastore2] 680c94e8-c6d9-4aab-92d4-047a0e70eec6 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.008934] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1781378-f4e4-4b71-be55-2737886b82fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.016993] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for the task: (returnval){ [ 1011.016993] env[63418]: value = "task-1245460" [ 1011.016993] env[63418]: _type = "Task" [ 1011.016993] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.025431] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245460, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.064604] env[63418]: DEBUG oslo_vmware.api [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245458, 'name': PowerOffVM_Task, 'duration_secs': 0.221425} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.065018] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.065281] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1011.065643] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7eba4c55-4015-46f8-9e92-35751814e7cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.195476] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.195728] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.196024] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore1] 9ae2cb1e-3999-4471-8a81-7f86db857ff3 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.196323] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b344858-9024-493d-9c8a-d04e4b76a08c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.204511] env[63418]: DEBUG oslo_vmware.api [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1011.204511] env[63418]: value = "task-1245462" [ 1011.204511] env[63418]: _type = "Task" [ 1011.204511] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.217687] env[63418]: DEBUG oslo_vmware.api [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.246591] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b3e2e1-bcb1-4d02-b552-4c4c57c12783 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.255303] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90907708-3229-453b-a413-deca6c41dcad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.287144] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8bdfcf-7399-4b01-a6df-2c3cc6bb0f7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.297669] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de488514-8ba8-4e00-a036-d2cf32c7c9c1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.320434] env[63418]: DEBUG nova.compute.provider_tree [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.527827] env[63418]: DEBUG oslo_vmware.api [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Task: {'id': task-1245460, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283874} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.529124] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.529124] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.529124] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.529124] env[63418]: INFO nova.compute.manager [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1011.529375] env[63418]: DEBUG oslo.service.loopingcall [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.529652] env[63418]: DEBUG nova.compute.manager [-] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1011.529743] env[63418]: DEBUG nova.network.neutron [-] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.717623] env[63418]: DEBUG oslo_vmware.api [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298638} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.717884] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.718040] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.718324] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.718537] env[63418]: INFO nova.compute.manager [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1011.718837] env[63418]: DEBUG oslo.service.loopingcall [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.719119] env[63418]: DEBUG nova.compute.manager [-] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1011.719239] env[63418]: DEBUG nova.network.neutron [-] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.790446] env[63418]: DEBUG nova.compute.manager [req-cac293d0-f6e9-4337-9fb8-aa4c71713f60 req-d06d02d3-e369-437d-bc6f-66546612006a service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Received event network-vif-deleted-492342f5-9181-4e37-9a04-ea9419e43523 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1011.791256] env[63418]: INFO nova.compute.manager [req-cac293d0-f6e9-4337-9fb8-aa4c71713f60 req-d06d02d3-e369-437d-bc6f-66546612006a service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Neutron deleted interface 492342f5-9181-4e37-9a04-ea9419e43523; detaching it from the instance and deleting it from the info cache [ 1011.791256] env[63418]: DEBUG nova.network.neutron [req-cac293d0-f6e9-4337-9fb8-aa4c71713f60 req-d06d02d3-e369-437d-bc6f-66546612006a service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.842573] env[63418]: ERROR nova.scheduler.client.report [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [req-13492c46-d586-4f55-8928-9adcb61a6946] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-13492c46-d586-4f55-8928-9adcb61a6946"}]} [ 1011.858327] env[63418]: DEBUG nova.scheduler.client.report [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1011.873137] env[63418]: DEBUG nova.scheduler.client.report [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1011.874071] env[63418]: DEBUG nova.compute.provider_tree [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.892069] env[63418]: DEBUG nova.scheduler.client.report [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1011.918086] env[63418]: DEBUG nova.scheduler.client.report [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1011.988735] env[63418]: DEBUG nova.compute.manager [req-f37fbe64-8a42-4116-a37d-e4aa7edfce5a req-954ea9e0-2b4e-4ec1-9dda-70a2d0a9fe21 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Received event network-vif-deleted-24cd6ab7-ed75-4c98-9839-27d80bd00464 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1011.989160] env[63418]: INFO nova.compute.manager [req-f37fbe64-8a42-4116-a37d-e4aa7edfce5a req-954ea9e0-2b4e-4ec1-9dda-70a2d0a9fe21 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Neutron deleted interface 24cd6ab7-ed75-4c98-9839-27d80bd00464; detaching it from the instance and deleting it from the info cache [ 1011.989509] env[63418]: DEBUG nova.network.neutron [req-f37fbe64-8a42-4116-a37d-e4aa7edfce5a req-954ea9e0-2b4e-4ec1-9dda-70a2d0a9fe21 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.075826] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aeeb789-b12c-416c-bf8f-60f6b96735fe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.087526] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a598963-9a7d-40d5-83f2-95bd97e520cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.125425] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d289c18-9d1c-4287-b488-457386679225 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.135044] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a25979-29a9-44e0-998f-fdd3772ad37d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.150515] env[63418]: DEBUG nova.compute.provider_tree [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.264656] env[63418]: DEBUG nova.network.neutron [-] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.297135] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52de109b-49a0-491b-9415-b13b7857aa83 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.307632] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655f70fd-9930-47af-9672-d1a5f92453cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.336921] env[63418]: DEBUG nova.compute.manager [req-cac293d0-f6e9-4337-9fb8-aa4c71713f60 req-d06d02d3-e369-437d-bc6f-66546612006a service nova] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Detach interface failed, port_id=492342f5-9181-4e37-9a04-ea9419e43523, reason: Instance 680c94e8-c6d9-4aab-92d4-047a0e70eec6 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1012.466987] env[63418]: DEBUG nova.network.neutron [-] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.492921] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76803aac-e663-425a-8f3b-45319c53b42e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.504056] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90c0b57-5b20-4f6f-9f69-e5837fbb8de7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.534396] env[63418]: DEBUG nova.compute.manager [req-f37fbe64-8a42-4116-a37d-e4aa7edfce5a req-954ea9e0-2b4e-4ec1-9dda-70a2d0a9fe21 service nova] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Detach interface failed, port_id=24cd6ab7-ed75-4c98-9839-27d80bd00464, reason: Instance 9ae2cb1e-3999-4471-8a81-7f86db857ff3 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1012.683647] env[63418]: DEBUG nova.scheduler.client.report [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1012.683826] env[63418]: DEBUG nova.compute.provider_tree [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 131 to 132 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1012.684105] env[63418]: DEBUG nova.compute.provider_tree [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.767336] env[63418]: INFO nova.compute.manager [-] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Took 1.24 seconds to deallocate network for instance. [ 1012.969612] env[63418]: INFO nova.compute.manager [-] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Took 1.25 seconds to deallocate network for instance. [ 1013.190273] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.112s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.190273] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1013.193200] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.759s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.193536] env[63418]: DEBUG nova.objects.instance [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lazy-loading 'resources' on Instance uuid 0963bddc-6d3d-424d-a218-a2341fa9dab7 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.276386] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.477274] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.698219] env[63418]: DEBUG nova.compute.utils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.703073] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1013.705031] env[63418]: DEBUG nova.network.neutron [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.754340] env[63418]: DEBUG nova.policy [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8425c7b68b8f4531a1b18f50354400e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '586d89911a3946e0a1fd9928017565f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.844206] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d648458-f9d8-491e-820a-867ccfdf2d8c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.852901] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fdfb5c-e9a1-4e5e-afb6-4c2ca9451196 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.889022] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3ade74-0665-4769-8de7-c63a5d74f48b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.899602] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8d4a6a-2f99-487f-8d6f-ade180790ffe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.915426] env[63418]: DEBUG nova.compute.provider_tree [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.048464] env[63418]: DEBUG nova.network.neutron [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Successfully created port: 0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.204248] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1014.419214] env[63418]: DEBUG nova.scheduler.client.report [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1014.924655] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.927035] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.651s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.927277] env[63418]: DEBUG nova.objects.instance [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lazy-loading 'resources' on Instance uuid 680c94e8-c6d9-4aab-92d4-047a0e70eec6 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.946668] env[63418]: INFO nova.scheduler.client.report [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Deleted allocations for instance 0963bddc-6d3d-424d-a218-a2341fa9dab7 [ 1015.216035] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1015.242130] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.242419] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.242575] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.242762] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.242916] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.243087] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.243318] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.243485] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.243659] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.243825] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.244007] env[63418]: DEBUG nova.virt.hardware [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.245112] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8e4c2d-7d80-44eb-b5fe-5eda6a561ed6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.253819] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9fe6de-d853-45b5-ba1f-254661044f24 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.455320] env[63418]: DEBUG oslo_concurrency.lockutils [None req-32c3ec58-6f5b-4fb7-98a1-f29c072161f2 tempest-ServerDiskConfigTestJSON-605256464 tempest-ServerDiskConfigTestJSON-605256464-project-member] Lock "0963bddc-6d3d-424d-a218-a2341fa9dab7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.707s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.556753] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941d68f6-9dba-4ae8-a273-b8d3c2a7d465 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.567638] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fb6174-e92b-4f8c-8529-619fad8e348b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.619363] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72b2c94-3caf-4d1d-a2bf-1d55003c31e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.630274] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b94802-5ceb-4d73-afdf-89df855a3917 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.647073] env[63418]: DEBUG nova.compute.provider_tree [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.659818] env[63418]: DEBUG nova.compute.manager [req-9e8a3ac4-fd79-4f93-92ed-24e44f5fcca2 req-3e715afe-79a5-4706-a07d-1bed2ff20478 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Received event network-vif-plugged-0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1015.660078] env[63418]: DEBUG oslo_concurrency.lockutils [req-9e8a3ac4-fd79-4f93-92ed-24e44f5fcca2 req-3e715afe-79a5-4706-a07d-1bed2ff20478 service nova] Acquiring lock "80777e38-728b-4d2d-9003-f55ea7999299-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.660350] env[63418]: DEBUG oslo_concurrency.lockutils [req-9e8a3ac4-fd79-4f93-92ed-24e44f5fcca2 req-3e715afe-79a5-4706-a07d-1bed2ff20478 service nova] Lock "80777e38-728b-4d2d-9003-f55ea7999299-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.660563] env[63418]: DEBUG oslo_concurrency.lockutils [req-9e8a3ac4-fd79-4f93-92ed-24e44f5fcca2 req-3e715afe-79a5-4706-a07d-1bed2ff20478 service nova] Lock "80777e38-728b-4d2d-9003-f55ea7999299-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.660771] env[63418]: DEBUG nova.compute.manager [req-9e8a3ac4-fd79-4f93-92ed-24e44f5fcca2 req-3e715afe-79a5-4706-a07d-1bed2ff20478 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] No waiting events found dispatching network-vif-plugged-0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1015.660969] env[63418]: WARNING nova.compute.manager [req-9e8a3ac4-fd79-4f93-92ed-24e44f5fcca2 req-3e715afe-79a5-4706-a07d-1bed2ff20478 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Received unexpected event network-vif-plugged-0079f927-f058-4093-b9d5-7950de3da99a for instance with vm_state building and task_state spawning. [ 1015.760844] env[63418]: DEBUG nova.network.neutron [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Successfully updated port: 0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.152223] env[63418]: DEBUG nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1016.264190] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.264367] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquired lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.264528] env[63418]: DEBUG nova.network.neutron [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.658802] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.732s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.661104] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.184s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.661345] env[63418]: DEBUG nova.objects.instance [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid 9ae2cb1e-3999-4471-8a81-7f86db857ff3 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.684078] env[63418]: INFO nova.scheduler.client.report [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Deleted allocations for instance 680c94e8-c6d9-4aab-92d4-047a0e70eec6 [ 1016.800262] env[63418]: DEBUG nova.network.neutron [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.952069] env[63418]: DEBUG nova.network.neutron [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updating instance_info_cache with network_info: [{"id": "0079f927-f058-4093-b9d5-7950de3da99a", "address": "fa:16:3e:33:db:92", "network": {"id": "cdc1990f-d683-417c-87e7-a3c0f63e90c4", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1469581719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "586d89911a3946e0a1fd9928017565f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0079f927-f0", "ovs_interfaceid": "0079f927-f058-4093-b9d5-7950de3da99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.193581] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4171cf57-cd0e-4915-b73d-49de8ef5539a tempest-DeleteServersTestJSON-1357947180 tempest-DeleteServersTestJSON-1357947180-project-member] Lock "680c94e8-c6d9-4aab-92d4-047a0e70eec6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.309s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.325525] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d516f44-3548-4ba6-b426-92f3d5704175 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.334103] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29e781b-6f6e-4954-8502-afb11bee23ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.371131] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a6078d-a31d-4409-919e-c2ba38f3d000 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.380342] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79e4522-9320-4fb9-9765-898f9eb78d46 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.398872] env[63418]: DEBUG nova.compute.provider_tree [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.455214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Releasing lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.455551] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Instance network_info: |[{"id": "0079f927-f058-4093-b9d5-7950de3da99a", "address": "fa:16:3e:33:db:92", "network": {"id": "cdc1990f-d683-417c-87e7-a3c0f63e90c4", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1469581719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "586d89911a3946e0a1fd9928017565f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0079f927-f0", "ovs_interfaceid": "0079f927-f058-4093-b9d5-7950de3da99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1017.456183] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:db:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0079f927-f058-4093-b9d5-7950de3da99a', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.463852] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Creating folder: Project (586d89911a3946e0a1fd9928017565f6). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1017.464191] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae3095cd-173e-47a1-b2f7-6b756683de4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.476864] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Created folder: Project (586d89911a3946e0a1fd9928017565f6) in parent group-v268354. [ 1017.477113] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Creating folder: Instances. Parent ref: group-v268531. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1017.477555] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c701fa15-d8c9-4daa-9f13-c5067eab751b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.490969] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Created folder: Instances in parent group-v268531. [ 1017.491456] env[63418]: DEBUG oslo.service.loopingcall [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.491784] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.492180] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce2fe4de-4ebc-432d-b666-efa9cd544dc5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.515175] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.515175] env[63418]: value = "task-1245466" [ 1017.515175] env[63418]: _type = "Task" [ 1017.515175] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.524401] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245466, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.691769] env[63418]: DEBUG nova.compute.manager [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Received event network-changed-0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1017.692019] env[63418]: DEBUG nova.compute.manager [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Refreshing instance network info cache due to event network-changed-0079f927-f058-4093-b9d5-7950de3da99a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1017.692292] env[63418]: DEBUG oslo_concurrency.lockutils [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] Acquiring lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.692461] env[63418]: DEBUG oslo_concurrency.lockutils [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] Acquired lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.692668] env[63418]: DEBUG nova.network.neutron [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Refreshing network info cache for port 0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.745840] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525832df-8534-18db-d8d2-53aee909f2f6/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1017.746798] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9758245d-3e02-45ce-b3f5-ed37e91ef4bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.754327] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525832df-8534-18db-d8d2-53aee909f2f6/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1017.754508] env[63418]: ERROR oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525832df-8534-18db-d8d2-53aee909f2f6/disk-0.vmdk due to incomplete transfer. [ 1017.754825] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cc2e9d6f-cb45-42db-a899-1d1ddb369519 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.764483] env[63418]: DEBUG oslo_vmware.rw_handles [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525832df-8534-18db-d8d2-53aee909f2f6/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1017.764755] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Uploaded image 8691809c-e40f-490b-bcba-aaf810e4f073 to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1017.767212] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1017.767470] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-92abf79a-5483-4637-a171-54d75aa101da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.774954] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1017.774954] env[63418]: value = "task-1245467" [ 1017.774954] env[63418]: _type = "Task" [ 1017.774954] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.784040] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245467, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.902549] env[63418]: DEBUG nova.scheduler.client.report [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1018.025644] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245466, 'name': CreateVM_Task, 'duration_secs': 0.36053} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.025644] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.026230] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.026401] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.026842] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.027010] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01b088c4-5fa3-4c58-8151-86996c257e58 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.032074] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1018.032074] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521c751c-74e8-e995-e9ca-f7e7b2ca60e2" [ 1018.032074] env[63418]: _type = "Task" [ 1018.032074] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.040382] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521c751c-74e8-e995-e9ca-f7e7b2ca60e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.286609] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245467, 'name': Destroy_Task, 'duration_secs': 0.373744} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.286917] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Destroyed the VM [ 1018.287125] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1018.287393] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c3d2596b-8313-4df5-a53b-7c8fba6dd4db {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.297617] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1018.297617] env[63418]: value = "task-1245468" [ 1018.297617] env[63418]: _type = "Task" [ 1018.297617] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.306551] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245468, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.407743] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.747s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.442709] env[63418]: INFO nova.scheduler.client.report [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance 9ae2cb1e-3999-4471-8a81-7f86db857ff3 [ 1018.510882] env[63418]: DEBUG nova.network.neutron [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updated VIF entry in instance network info cache for port 0079f927-f058-4093-b9d5-7950de3da99a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.511343] env[63418]: DEBUG nova.network.neutron [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updating instance_info_cache with network_info: [{"id": "0079f927-f058-4093-b9d5-7950de3da99a", "address": "fa:16:3e:33:db:92", "network": {"id": "cdc1990f-d683-417c-87e7-a3c0f63e90c4", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1469581719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "586d89911a3946e0a1fd9928017565f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0079f927-f0", "ovs_interfaceid": "0079f927-f058-4093-b9d5-7950de3da99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.543548] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521c751c-74e8-e995-e9ca-f7e7b2ca60e2, 'name': SearchDatastore_Task, 'duration_secs': 0.012512} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.543889] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.544150] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.544398] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.544547] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.544963] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.545356] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e63b4d37-65c9-457c-b6be-77d9b9be7df6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.559303] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.559664] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.560850] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ac88593-362e-44d7-af46-54ffb1da3daf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.569172] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1018.569172] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a48056-3292-ff87-6431-b7caef52e3c8" [ 1018.569172] env[63418]: _type = "Task" [ 1018.569172] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.580162] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a48056-3292-ff87-6431-b7caef52e3c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.811241] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245468, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.955039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-aa26aa6d-3514-44f1-aade-14892f0d7354 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "9ae2cb1e-3999-4471-8a81-7f86db857ff3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.932s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.015682] env[63418]: DEBUG oslo_concurrency.lockutils [req-f2e57196-833b-4bb4-a2a5-db360af703f9 req-912ca480-58cc-4e08-9517-c20e6f9f7140 service nova] Releasing lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.080659] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a48056-3292-ff87-6431-b7caef52e3c8, 'name': SearchDatastore_Task, 'duration_secs': 0.012726} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.081659] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5acc46b3-de4d-4b30-8fbf-e1d07874a65f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.088066] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1019.088066] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]526b6449-28f9-5271-ac4d-b265a3c055ac" [ 1019.088066] env[63418]: _type = "Task" [ 1019.088066] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.099530] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526b6449-28f9-5271-ac4d-b265a3c055ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.312490] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245468, 'name': RemoveSnapshot_Task, 'duration_secs': 0.553029} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.312921] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1019.313213] env[63418]: DEBUG nova.compute.manager [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1019.314051] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d93a691-15b5-49da-ba26-c585b88663af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.603150] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]526b6449-28f9-5271-ac4d-b265a3c055ac, 'name': SearchDatastore_Task, 'duration_secs': 0.020622} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.603150] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.603485] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 80777e38-728b-4d2d-9003-f55ea7999299/80777e38-728b-4d2d-9003-f55ea7999299.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.603582] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b921c05-4a14-4077-b283-974780f1ab5d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.613135] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1019.613135] env[63418]: value = "task-1245469" [ 1019.613135] env[63418]: _type = "Task" [ 1019.613135] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.626659] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245469, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.828423] env[63418]: INFO nova.compute.manager [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Shelve offloading [ 1020.021684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.022323] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.022323] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.022508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.022693] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.024979] env[63418]: INFO nova.compute.manager [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Terminating instance [ 1020.126860] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245469, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.334115] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.334507] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-869bb032-ac46-40ca-83ae-1db93ff257d3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.349215] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1020.349215] env[63418]: value = "task-1245470" [ 1020.349215] env[63418]: _type = "Task" [ 1020.349215] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.366187] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1020.366487] env[63418]: DEBUG nova.compute.manager [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1020.367377] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c8c114-7ad3-4329-89f7-8562ef4060d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.377490] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.377689] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.377819] env[63418]: DEBUG nova.network.neutron [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.529279] env[63418]: DEBUG nova.compute.manager [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1020.529279] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.529975] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31a8d2a-64a3-48cf-8660-63d55eeda41d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.539495] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.539838] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ccfac12-506d-42d8-a374-e621214aae3c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.547671] env[63418]: DEBUG oslo_vmware.api [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1020.547671] env[63418]: value = "task-1245471" [ 1020.547671] env[63418]: _type = "Task" [ 1020.547671] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.559186] env[63418]: DEBUG oslo_vmware.api [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.627758] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245469, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.813764} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.628072] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 80777e38-728b-4d2d-9003-f55ea7999299/80777e38-728b-4d2d-9003-f55ea7999299.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.628299] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.628562] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8571c4b5-017f-4902-93ef-a4f26a384656 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.637054] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1020.637054] env[63418]: value = "task-1245472" [ 1020.637054] env[63418]: _type = "Task" [ 1020.637054] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.646991] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245472, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.061018] env[63418]: DEBUG oslo_vmware.api [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245471, 'name': PowerOffVM_Task, 'duration_secs': 0.228253} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.061018] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.061018] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.061018] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05f3550e-d04f-40ee-897a-a31fbf941964 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.148927] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245472, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113414} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.149111] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.149958] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e4f11a-6855-4420-847b-2722eb2900ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.157502] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.157743] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.158014] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore1] 3e645774-d1b3-468b-86a5-ff0c07e77f4e {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.158325] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82f8fe4a-dbc1-49a8-b3e1-13d7a169b5b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.181463] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 80777e38-728b-4d2d-9003-f55ea7999299/80777e38-728b-4d2d-9003-f55ea7999299.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.181463] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e397bcdf-9b90-4024-9b42-83666e66498c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.200252] env[63418]: DEBUG oslo_vmware.api [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1021.200252] env[63418]: value = "task-1245474" [ 1021.200252] env[63418]: _type = "Task" [ 1021.200252] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.209162] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1021.209162] env[63418]: value = "task-1245475" [ 1021.209162] env[63418]: _type = "Task" [ 1021.209162] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.217344] env[63418]: DEBUG oslo_vmware.api [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.223565] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245475, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.255385] env[63418]: DEBUG nova.network.neutron [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.299555] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "d0633a7d-33b4-4613-9501-81cef2ab3680" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.300148] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.329597] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.329853] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.715483] env[63418]: DEBUG oslo_vmware.api [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263822} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.719359] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.722349] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.722349] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.722349] env[63418]: INFO nova.compute.manager [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1021.722349] env[63418]: DEBUG oslo.service.loopingcall [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.722349] env[63418]: DEBUG nova.compute.manager [-] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1021.722349] env[63418]: DEBUG nova.network.neutron [-] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.729251] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.760512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.803149] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1021.832703] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1022.083275] env[63418]: DEBUG nova.compute.manager [req-8f82def4-0a62-4d8c-b0e5-0e78e9d83658 req-73d8881f-707a-4314-a3f4-f9baa22574a3 service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Received event network-vif-deleted-7281180f-8ed9-4bba-a220-c22a98d334c7 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1022.083275] env[63418]: INFO nova.compute.manager [req-8f82def4-0a62-4d8c-b0e5-0e78e9d83658 req-73d8881f-707a-4314-a3f4-f9baa22574a3 service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Neutron deleted interface 7281180f-8ed9-4bba-a220-c22a98d334c7; detaching it from the instance and deleting it from the info cache [ 1022.083275] env[63418]: DEBUG nova.network.neutron [req-8f82def4-0a62-4d8c-b0e5-0e78e9d83658 req-73d8881f-707a-4314-a3f4-f9baa22574a3 service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.131766] env[63418]: DEBUG nova.compute.manager [req-8859c8af-eb69-4be7-ab05-f08235bc8cef req-6ca6fadd-34d3-4c46-a7b1-e623bfa9244e service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-vif-unplugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1022.132207] env[63418]: DEBUG oslo_concurrency.lockutils [req-8859c8af-eb69-4be7-ab05-f08235bc8cef req-6ca6fadd-34d3-4c46-a7b1-e623bfa9244e service nova] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.132556] env[63418]: DEBUG oslo_concurrency.lockutils [req-8859c8af-eb69-4be7-ab05-f08235bc8cef req-6ca6fadd-34d3-4c46-a7b1-e623bfa9244e service nova] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.132885] env[63418]: DEBUG oslo_concurrency.lockutils [req-8859c8af-eb69-4be7-ab05-f08235bc8cef req-6ca6fadd-34d3-4c46-a7b1-e623bfa9244e service nova] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.133224] env[63418]: DEBUG nova.compute.manager [req-8859c8af-eb69-4be7-ab05-f08235bc8cef req-6ca6fadd-34d3-4c46-a7b1-e623bfa9244e service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] No waiting events found dispatching network-vif-unplugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1022.133543] env[63418]: WARNING nova.compute.manager [req-8859c8af-eb69-4be7-ab05-f08235bc8cef req-6ca6fadd-34d3-4c46-a7b1-e623bfa9244e service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received unexpected event network-vif-unplugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 for instance with vm_state shelved and task_state shelving_offloading. [ 1022.228409] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245475, 'name': ReconfigVM_Task, 'duration_secs': 0.580063} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.228409] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 80777e38-728b-4d2d-9003-f55ea7999299/80777e38-728b-4d2d-9003-f55ea7999299.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.228409] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a371c9a2-b234-41fd-8313-958a1851196f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.236262] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1022.236262] env[63418]: value = "task-1245476" [ 1022.236262] env[63418]: _type = "Task" [ 1022.236262] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.247347] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245476, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.249649] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.250428] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e420ce1a-10e9-4054-994d-3b4c963743b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.258260] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.260089] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2242120f-7e66-430b-aea8-217ad6894132 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.332156] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.332542] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.334469] env[63418]: INFO nova.compute.claims [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.340873] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.341550] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.341812] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleting the datastore file [datastore2] d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.344117] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc131031-e8de-49f3-9789-6a82c2380772 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.353021] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1022.353021] env[63418]: value = "task-1245478" [ 1022.353021] env[63418]: _type = "Task" [ 1022.353021] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.363952] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.365420] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.540229] env[63418]: DEBUG nova.network.neutron [-] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.587979] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d64720c1-1d54-4567-b79e-3c585c56dfe7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.599950] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17724a5a-669f-46f2-9528-f82c9725296c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.628773] env[63418]: DEBUG nova.compute.manager [req-8f82def4-0a62-4d8c-b0e5-0e78e9d83658 req-73d8881f-707a-4314-a3f4-f9baa22574a3 service nova] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Detach interface failed, port_id=7281180f-8ed9-4bba-a220-c22a98d334c7, reason: Instance 3e645774-d1b3-468b-86a5-ff0c07e77f4e could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1022.745424] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245476, 'name': Rename_Task, 'duration_secs': 0.17841} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.745822] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.746168] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff492805-8cbc-46b6-85be-ee5703ad061a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.754371] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1022.754371] env[63418]: value = "task-1245479" [ 1022.754371] env[63418]: _type = "Task" [ 1022.754371] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.762738] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.862260] env[63418]: DEBUG oslo_vmware.api [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14971} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.862612] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.862871] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.863094] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.884144] env[63418]: INFO nova.scheduler.client.report [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted allocations for instance d52db46b-2461-4bd0-be57-d414250aac7e [ 1023.043203] env[63418]: INFO nova.compute.manager [-] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Took 1.32 seconds to deallocate network for instance. [ 1023.265540] env[63418]: DEBUG oslo_vmware.api [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245479, 'name': PowerOnVM_Task, 'duration_secs': 0.484256} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.265882] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.266109] env[63418]: INFO nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Took 8.05 seconds to spawn the instance on the hypervisor. [ 1023.266297] env[63418]: DEBUG nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1023.267080] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7da648-3447-4c99-9441-82a82ec8d9fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.389395] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.444712] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d548dc0f-299c-4957-a6bf-080cbcaaa8f1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.452687] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3187e28-aaf9-477c-9743-9055a66c9817 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.483241] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deeb1c65-11c0-4b85-b06e-481b66c09f5a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.491074] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c0d732-664c-49ba-a836-889aaf204a8b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.504056] env[63418]: DEBUG nova.compute.provider_tree [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.550100] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.785742] env[63418]: INFO nova.compute.manager [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Took 13.73 seconds to build instance. [ 1024.007506] env[63418]: DEBUG nova.scheduler.client.report [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1024.155894] env[63418]: DEBUG nova.compute.manager [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1024.156082] env[63418]: DEBUG nova.compute.manager [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing instance network info cache due to event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1024.156308] env[63418]: DEBUG oslo_concurrency.lockutils [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.156452] env[63418]: DEBUG oslo_concurrency.lockutils [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.156613] env[63418]: DEBUG nova.network.neutron [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.288205] env[63418]: DEBUG oslo_concurrency.lockutils [None req-252f0f7f-85c8-48aa-a941-69edb05f7eaf tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "80777e38-728b-4d2d-9003-f55ea7999299" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.238s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.513123] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.180s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.513760] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1024.516539] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.151s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.517818] env[63418]: INFO nova.compute.claims [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.021685] env[63418]: DEBUG nova.compute.utils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.029228] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1025.029421] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1025.079304] env[63418]: DEBUG nova.policy [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc5a12be9ce64b5294d8aab8af52e61c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c585aa2fd2142aba53f8d5a91f5e332', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1025.309711] env[63418]: DEBUG nova.network.neutron [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updated VIF entry in instance network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.310101] env[63418]: DEBUG nova.network.neutron [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.392690] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Successfully created port: a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.536120] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1025.685425] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d90b31c-9382-4f5a-ba04-81ca3f38bded {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.693929] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c67f9e-4628-49b8-bc6d-5c4519887486 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.724338] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcc2533-360b-4bd5-ad8a-356c9445b261 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.732541] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5f7594-0652-49ef-bf3c-7f060697416d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.747984] env[63418]: DEBUG nova.compute.provider_tree [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.812727] env[63418]: DEBUG oslo_concurrency.lockutils [req-7041cea2-79f3-4d14-adf7-80ebd030f641 req-7523fa98-2caa-4bb6-81a1-f898bd4f713f service nova] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.165771] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.180280] env[63418]: DEBUG nova.compute.manager [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Received event network-changed-0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1026.180527] env[63418]: DEBUG nova.compute.manager [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Refreshing instance network info cache due to event network-changed-0079f927-f058-4093-b9d5-7950de3da99a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1026.180793] env[63418]: DEBUG oslo_concurrency.lockutils [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] Acquiring lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.180962] env[63418]: DEBUG oslo_concurrency.lockutils [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] Acquired lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.181288] env[63418]: DEBUG nova.network.neutron [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Refreshing network info cache for port 0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.251010] env[63418]: DEBUG nova.scheduler.client.report [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1026.547059] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1026.572610] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.572900] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.573097] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.573310] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.573469] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.573647] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.573881] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.574118] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.574341] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.574523] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.574775] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.575758] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0d1d5f-ce26-4d58-891d-7ce640a22abb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.586123] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd247d3-b209-451d-8833-ab8a8332c1cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.755604] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.755974] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1026.758450] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.369s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.758710] env[63418]: DEBUG nova.objects.instance [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'resources' on Instance uuid d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.794122] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.794300] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.842238] env[63418]: DEBUG nova.compute.manager [req-4f0e689e-dc2c-4b1b-b2e7-20440048235c req-abbe39c7-2d47-459b-a1c6-29652c594e13 service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Received event network-vif-plugged-a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1026.842747] env[63418]: DEBUG oslo_concurrency.lockutils [req-4f0e689e-dc2c-4b1b-b2e7-20440048235c req-abbe39c7-2d47-459b-a1c6-29652c594e13 service nova] Acquiring lock "d0633a7d-33b4-4613-9501-81cef2ab3680-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.842747] env[63418]: DEBUG oslo_concurrency.lockutils [req-4f0e689e-dc2c-4b1b-b2e7-20440048235c req-abbe39c7-2d47-459b-a1c6-29652c594e13 service nova] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.842939] env[63418]: DEBUG oslo_concurrency.lockutils [req-4f0e689e-dc2c-4b1b-b2e7-20440048235c req-abbe39c7-2d47-459b-a1c6-29652c594e13 service nova] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.842939] env[63418]: DEBUG nova.compute.manager [req-4f0e689e-dc2c-4b1b-b2e7-20440048235c req-abbe39c7-2d47-459b-a1c6-29652c594e13 service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] No waiting events found dispatching network-vif-plugged-a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1026.843106] env[63418]: WARNING nova.compute.manager [req-4f0e689e-dc2c-4b1b-b2e7-20440048235c req-abbe39c7-2d47-459b-a1c6-29652c594e13 service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Received unexpected event network-vif-plugged-a5d70b5d-10c4-4709-8fe1-f84cf0e84489 for instance with vm_state building and task_state spawning. [ 1026.906698] env[63418]: DEBUG nova.network.neutron [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updated VIF entry in instance network info cache for port 0079f927-f058-4093-b9d5-7950de3da99a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1026.906812] env[63418]: DEBUG nova.network.neutron [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updating instance_info_cache with network_info: [{"id": "0079f927-f058-4093-b9d5-7950de3da99a", "address": "fa:16:3e:33:db:92", "network": {"id": "cdc1990f-d683-417c-87e7-a3c0f63e90c4", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1469581719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "586d89911a3946e0a1fd9928017565f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0079f927-f0", "ovs_interfaceid": "0079f927-f058-4093-b9d5-7950de3da99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.926547] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Successfully updated port: a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.261292] env[63418]: DEBUG nova.compute.utils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.262881] env[63418]: DEBUG nova.objects.instance [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'numa_topology' on Instance uuid d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.264680] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1027.265307] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1027.297061] env[63418]: INFO nova.compute.manager [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Detaching volume e5151ce5-c6ee-4092-9eeb-e00affc031dd [ 1027.305446] env[63418]: DEBUG nova.policy [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc5a12be9ce64b5294d8aab8af52e61c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c585aa2fd2142aba53f8d5a91f5e332', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.328847] env[63418]: INFO nova.virt.block_device [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Attempting to driver detach volume e5151ce5-c6ee-4092-9eeb-e00affc031dd from mountpoint /dev/sdb [ 1027.329103] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1027.329294] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268519', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'name': 'volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcbaf4f-ee24-4072-83ae-ffde59478928', 'attached_at': '', 'detached_at': '', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'serial': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1027.330173] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941c8e6a-041c-4e59-83fa-a02507b5f835 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.352814] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f50469f-fc07-448e-9c56-8adffdd90743 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.360907] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14ce68c-4bf7-4544-8f12-ab4e3deb607d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.382199] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f44125-b03c-4b0b-ad6e-5dbe1780201e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.397849] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] The volume has not been displaced from its original location: [datastore2] volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd/volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1027.403429] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1027.403681] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e62a303a-2b73-4674-a178-afda803c6541 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.416672] env[63418]: DEBUG oslo_concurrency.lockutils [req-acb70ee9-9f52-47ff-a99a-a02802734535 req-57b29ba1-1b6e-4978-a993-6a6c8f1b9e7d service nova] Releasing lock "refresh_cache-80777e38-728b-4d2d-9003-f55ea7999299" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.424676] env[63418]: DEBUG oslo_vmware.api [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1027.424676] env[63418]: value = "task-1245480" [ 1027.424676] env[63418]: _type = "Task" [ 1027.424676] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.433453] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "refresh_cache-d0633a7d-33b4-4613-9501-81cef2ab3680" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.433639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "refresh_cache-d0633a7d-33b4-4613-9501-81cef2ab3680" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.433741] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.434952] env[63418]: DEBUG oslo_vmware.api [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245480, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.555546] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Successfully created port: a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.765376] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1027.768998] env[63418]: DEBUG nova.objects.base [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.884994] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70517a60-cbdf-44c3-9692-1cbea448b677 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.894489] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718493a1-85bd-4455-b6aa-24e93f4678be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.929737] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e74b11-e22d-45cc-b171-13d3cc241992 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.941644] env[63418]: DEBUG oslo_vmware.api [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245480, 'name': ReconfigVM_Task, 'duration_secs': 0.238316} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.942951] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2dc30d4-a35c-4fcc-9ee6-e0952d72925b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.947781] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1027.952934] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f425ef87-d44e-4b49-9800-33141cddc0d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.974220] env[63418]: DEBUG nova.compute.provider_tree [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.977625] env[63418]: DEBUG oslo_vmware.api [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1027.977625] env[63418]: value = "task-1245481" [ 1027.977625] env[63418]: _type = "Task" [ 1027.977625] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.982753] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.988871] env[63418]: DEBUG oslo_vmware.api [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245481, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.123263] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Updating instance_info_cache with network_info: [{"id": "a5d70b5d-10c4-4709-8fe1-f84cf0e84489", "address": "fa:16:3e:e0:6c:d4", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5d70b5d-10", "ovs_interfaceid": "a5d70b5d-10c4-4709-8fe1-f84cf0e84489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.478628] env[63418]: DEBUG nova.scheduler.client.report [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1028.491720] env[63418]: DEBUG oslo_vmware.api [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245481, 'name': ReconfigVM_Task, 'duration_secs': 0.143451} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.492063] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268519', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'name': 'volume-e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcbaf4f-ee24-4072-83ae-ffde59478928', 'attached_at': '', 'detached_at': '', 'volume_id': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd', 'serial': 'e5151ce5-c6ee-4092-9eeb-e00affc031dd'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1028.625348] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "refresh_cache-d0633a7d-33b4-4613-9501-81cef2ab3680" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.625674] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Instance network_info: |[{"id": "a5d70b5d-10c4-4709-8fe1-f84cf0e84489", "address": "fa:16:3e:e0:6c:d4", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5d70b5d-10", "ovs_interfaceid": "a5d70b5d-10c4-4709-8fe1-f84cf0e84489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1028.626141] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:6c:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5d70b5d-10c4-4709-8fe1-f84cf0e84489', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.633656] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Creating folder: Project (4c585aa2fd2142aba53f8d5a91f5e332). Parent ref: group-v268354. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.633936] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-491a8d9c-ac0b-4e10-a60c-bb552ebad151 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.647542] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Created folder: Project (4c585aa2fd2142aba53f8d5a91f5e332) in parent group-v268354. [ 1028.647542] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Creating folder: Instances. Parent ref: group-v268534. {{(pid=63418) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.647698] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee871f88-ff61-49e0-bdc3-4e32817e9e7c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.657171] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Created folder: Instances in parent group-v268534. [ 1028.657405] env[63418]: DEBUG oslo.service.loopingcall [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.657600] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.657801] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2e7cd97-b8d9-4c10-9058-774941ceeb0e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.679259] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.679259] env[63418]: value = "task-1245484" [ 1028.679259] env[63418]: _type = "Task" [ 1028.679259] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.686819] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245484, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.781141] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1028.808750] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.808891] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.809062] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.809261] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.809411] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.809561] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.809772] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.809939] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.810128] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.810379] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.810467] env[63418]: DEBUG nova.virt.hardware [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.811347] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cbf7ac-2a47-43f5-81d1-b6ff33b5f11e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.820647] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe20971-0bb0-4597-a863-bc04c21eeb8e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.868290] env[63418]: DEBUG nova.compute.manager [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Received event network-changed-a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1028.868669] env[63418]: DEBUG nova.compute.manager [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Refreshing instance network info cache due to event network-changed-a5d70b5d-10c4-4709-8fe1-f84cf0e84489. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1028.868884] env[63418]: DEBUG oslo_concurrency.lockutils [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] Acquiring lock "refresh_cache-d0633a7d-33b4-4613-9501-81cef2ab3680" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.869080] env[63418]: DEBUG oslo_concurrency.lockutils [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] Acquired lock "refresh_cache-d0633a7d-33b4-4613-9501-81cef2ab3680" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.869287] env[63418]: DEBUG nova.network.neutron [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Refreshing network info cache for port a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.872077] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.872339] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.987070] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.228s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.993071] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.440s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.993392] env[63418]: DEBUG nova.objects.instance [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid 3e645774-d1b3-468b-86a5-ff0c07e77f4e {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.035793] env[63418]: DEBUG nova.objects.instance [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.190617] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245484, 'name': CreateVM_Task, 'duration_secs': 0.334474} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.190617] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1029.190816] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.190977] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.191311] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1029.191617] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b4623a2-099b-4b2a-8106-9f27548cc055 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.197127] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1029.197127] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529546c2-8956-0c6f-872e-72879682ab88" [ 1029.197127] env[63418]: _type = "Task" [ 1029.197127] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.205825] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529546c2-8956-0c6f-872e-72879682ab88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.243918] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Successfully updated port: a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.375619] env[63418]: DEBUG nova.compute.utils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.503498] env[63418]: DEBUG oslo_concurrency.lockutils [None req-b49620ef-61f0-4eb4-bfee-1f4ee04c5df7 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.128s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.505886] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.339s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.505886] env[63418]: INFO nova.compute.manager [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Unshelving [ 1029.581276] env[63418]: DEBUG nova.network.neutron [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Updated VIF entry in instance network info cache for port a5d70b5d-10c4-4709-8fe1-f84cf0e84489. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.581628] env[63418]: DEBUG nova.network.neutron [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Updating instance_info_cache with network_info: [{"id": "a5d70b5d-10c4-4709-8fe1-f84cf0e84489", "address": "fa:16:3e:e0:6c:d4", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5d70b5d-10", "ovs_interfaceid": "a5d70b5d-10c4-4709-8fe1-f84cf0e84489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.621562] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a48ac6-367f-429e-ae88-2060768c7b9f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.629875] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b775f4-b4d0-4671-b6bd-3a563eb68856 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.662646] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac61a5c-c549-4575-a8e2-3e800e975d87 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.671918] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed5eabd-354c-41e3-b966-fe7788de1268 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.685260] env[63418]: DEBUG nova.compute.provider_tree [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.706223] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529546c2-8956-0c6f-872e-72879682ab88, 'name': SearchDatastore_Task, 'duration_secs': 0.01046} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.706539] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.706774] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.707017] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.707183] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.707371] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.707617] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ff35982-5589-494c-88cb-16e6756186ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.716613] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.716745] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.717468] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-749385b5-a050-47c3-853b-2ceed58c440d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.722321] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1029.722321] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5278a8e3-9304-1731-9f2c-ed9ff1fd178b" [ 1029.722321] env[63418]: _type = "Task" [ 1029.722321] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.733402] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5278a8e3-9304-1731-9f2c-ed9ff1fd178b, 'name': SearchDatastore_Task, 'duration_secs': 0.008516} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.734124] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4928e3d-99ef-4fe0-90f2-b60bee584efd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.738998] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1029.738998] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256e109-fd98-ba29-416d-e95b2c754641" [ 1029.738998] env[63418]: _type = "Task" [ 1029.738998] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.746850] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "refresh_cache-2dae6f5b-9bdc-4862-8654-2684ea22c6ac" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.746987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "refresh_cache-2dae6f5b-9bdc-4862-8654-2684ea22c6ac" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.747152] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.749167] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256e109-fd98-ba29-416d-e95b2c754641, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.878821] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.985036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.048518] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2beaf8de-2309-44b1-9ea2-3ab4df9100ae tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.254s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.050221] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.066s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.050409] env[63418]: DEBUG nova.compute.manager [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1030.051654] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e8a30c-3669-4d55-82db-61547cf55288 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.059407] env[63418]: DEBUG nova.compute.manager [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1030.059947] env[63418]: DEBUG nova.objects.instance [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.084394] env[63418]: DEBUG oslo_concurrency.lockutils [req-85ef571d-1d21-487c-aafb-121e46fee611 req-964df92e-6e64-409e-9283-37f34b4a4c0c service nova] Releasing lock "refresh_cache-d0633a7d-33b4-4613-9501-81cef2ab3680" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.188666] env[63418]: DEBUG nova.scheduler.client.report [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1030.252615] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5256e109-fd98-ba29-416d-e95b2c754641, 'name': SearchDatastore_Task, 'duration_secs': 0.007948} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.252615] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.252732] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] d0633a7d-33b4-4613-9501-81cef2ab3680/d0633a7d-33b4-4613-9501-81cef2ab3680.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.253292] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78be2d89-99f8-474d-89e4-4a8311cc664c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.260225] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1030.260225] env[63418]: value = "task-1245485" [ 1030.260225] env[63418]: _type = "Task" [ 1030.260225] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.268944] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245485, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.280595] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.409382] env[63418]: DEBUG nova.network.neutron [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Updating instance_info_cache with network_info: [{"id": "a5c86e79-a127-4e43-9d97-8c69ea66352a", "address": "fa:16:3e:c8:14:5a", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5c86e79-a1", "ovs_interfaceid": "a5c86e79-a127-4e43-9d97-8c69ea66352a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.537208] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.693731] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.696786] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.159s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.696786] env[63418]: DEBUG nova.objects.instance [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'pci_requests' on Instance uuid d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.717631] env[63418]: INFO nova.scheduler.client.report [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance 3e645774-d1b3-468b-86a5-ff0c07e77f4e [ 1030.772911] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245485, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.896332] env[63418]: DEBUG nova.compute.manager [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Received event network-vif-plugged-a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1030.896589] env[63418]: DEBUG oslo_concurrency.lockutils [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] Acquiring lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.896864] env[63418]: DEBUG oslo_concurrency.lockutils [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.897140] env[63418]: DEBUG oslo_concurrency.lockutils [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.897356] env[63418]: DEBUG nova.compute.manager [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] No waiting events found dispatching network-vif-plugged-a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1030.897545] env[63418]: WARNING nova.compute.manager [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Received unexpected event network-vif-plugged-a5c86e79-a127-4e43-9d97-8c69ea66352a for instance with vm_state building and task_state spawning. [ 1030.897758] env[63418]: DEBUG nova.compute.manager [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Received event network-changed-a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1030.897970] env[63418]: DEBUG nova.compute.manager [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Refreshing instance network info cache due to event network-changed-a5c86e79-a127-4e43-9d97-8c69ea66352a. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1030.898176] env[63418]: DEBUG oslo_concurrency.lockutils [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] Acquiring lock "refresh_cache-2dae6f5b-9bdc-4862-8654-2684ea22c6ac" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.912348] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "refresh_cache-2dae6f5b-9bdc-4862-8654-2684ea22c6ac" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.912694] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Instance network_info: |[{"id": "a5c86e79-a127-4e43-9d97-8c69ea66352a", "address": "fa:16:3e:c8:14:5a", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5c86e79-a1", "ovs_interfaceid": "a5c86e79-a127-4e43-9d97-8c69ea66352a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1030.913068] env[63418]: DEBUG oslo_concurrency.lockutils [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] Acquired lock "refresh_cache-2dae6f5b-9bdc-4862-8654-2684ea22c6ac" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.913315] env[63418]: DEBUG nova.network.neutron [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Refreshing network info cache for port a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.914720] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:14:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5c86e79-a127-4e43-9d97-8c69ea66352a', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.922355] env[63418]: DEBUG oslo.service.loopingcall [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.925371] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.926125] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e26a53e6-c1fe-4cbc-a93c-5393a02fba0f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.942561] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.942993] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.943278] env[63418]: INFO nova.compute.manager [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Attaching volume 45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6 to /dev/sdb [ 1030.954240] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.954240] env[63418]: value = "task-1245486" [ 1030.954240] env[63418]: _type = "Task" [ 1030.954240] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.963607] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245486, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.983403] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7227384d-e28f-4c51-a285-24623423d5aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.991364] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22936a4-2ad5-442c-baf2-1611b3ca47fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.007708] env[63418]: DEBUG nova.virt.block_device [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating existing volume attachment record: 646c8c77-7020-4113-840a-2bc7f20e2fe4 {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1031.067957] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.068328] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-806eeb00-b0c2-43c3-9d9c-9ace7cf183c8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.077586] env[63418]: DEBUG oslo_vmware.api [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1031.077586] env[63418]: value = "task-1245487" [ 1031.077586] env[63418]: _type = "Task" [ 1031.077586] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.086744] env[63418]: DEBUG oslo_vmware.api [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.170342] env[63418]: DEBUG nova.network.neutron [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Updated VIF entry in instance network info cache for port a5c86e79-a127-4e43-9d97-8c69ea66352a. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.170889] env[63418]: DEBUG nova.network.neutron [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Updating instance_info_cache with network_info: [{"id": "a5c86e79-a127-4e43-9d97-8c69ea66352a", "address": "fa:16:3e:c8:14:5a", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5c86e79-a1", "ovs_interfaceid": "a5c86e79-a127-4e43-9d97-8c69ea66352a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.200606] env[63418]: DEBUG nova.objects.instance [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'numa_topology' on Instance uuid d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.226015] env[63418]: DEBUG oslo_concurrency.lockutils [None req-774ad7d1-949c-4f25-b700-e1fa96c0364c tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "3e645774-d1b3-468b-86a5-ff0c07e77f4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.204s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.271224] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245485, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542788} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.271547] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] d0633a7d-33b4-4613-9501-81cef2ab3680/d0633a7d-33b4-4613-9501-81cef2ab3680.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.271802] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.272101] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-271515d7-263d-425b-bca7-e2094370635e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.279594] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1031.279594] env[63418]: value = "task-1245491" [ 1031.279594] env[63418]: _type = "Task" [ 1031.279594] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.289502] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.465167] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245486, 'name': CreateVM_Task, 'duration_secs': 0.399783} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.465339] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.466081] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.466263] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.466597] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.466861] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0969371-ca15-48a3-a1d9-cedd593e940e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.472186] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1031.472186] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52dbf4a1-b030-e4af-ff51-af67e1d71abf" [ 1031.472186] env[63418]: _type = "Task" [ 1031.472186] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.481883] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52dbf4a1-b030-e4af-ff51-af67e1d71abf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.587273] env[63418]: DEBUG oslo_vmware.api [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245487, 'name': PowerOffVM_Task, 'duration_secs': 0.190377} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.587565] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.587764] env[63418]: DEBUG nova.compute.manager [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1031.588564] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf52bcc3-e445-452d-a8bf-85d054eb26ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.674130] env[63418]: DEBUG oslo_concurrency.lockutils [req-761e32b5-4682-4fec-9eb1-266344cfedcd req-b12d51cc-4811-4484-8229-39f4e752c2ee service nova] Releasing lock "refresh_cache-2dae6f5b-9bdc-4862-8654-2684ea22c6ac" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.703782] env[63418]: INFO nova.compute.claims [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.789599] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071675} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.789831] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.790611] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b057a03d-c967-4693-a864-13d8d718f222 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.814437] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] d0633a7d-33b4-4613-9501-81cef2ab3680/d0633a7d-33b4-4613-9501-81cef2ab3680.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.814729] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf84e024-af80-483f-85e2-a929d98f4a93 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.835023] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1031.835023] env[63418]: value = "task-1245492" [ 1031.835023] env[63418]: _type = "Task" [ 1031.835023] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.843220] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245492, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.982547] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52dbf4a1-b030-e4af-ff51-af67e1d71abf, 'name': SearchDatastore_Task, 'duration_secs': 0.010425} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.982865] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.983111] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.983446] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.983540] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.983720] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.983981] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac9afd6c-bbeb-4276-8bb8-cd051cfbafa2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.992324] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.992500] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.993208] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f341b6c6-721f-45b6-9654-fd44c1f6e3e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.998776] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1031.998776] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e56415-a817-05c5-3e69-fb8edf91e138" [ 1031.998776] env[63418]: _type = "Task" [ 1031.998776] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.006529] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e56415-a817-05c5-3e69-fb8edf91e138, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.100702] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d1dd40e-6e9c-47d8-984d-d01d1cc7bf05 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.050s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.345744] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245492, 'name': ReconfigVM_Task, 'duration_secs': 0.299117} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.346033] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Reconfigured VM instance instance-00000067 to attach disk [datastore2] d0633a7d-33b4-4613-9501-81cef2ab3680/d0633a7d-33b4-4613-9501-81cef2ab3680.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.346668] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f3837b0-8d83-493a-b646-c44c6e1ef3ca {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.353963] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1032.353963] env[63418]: value = "task-1245493" [ 1032.353963] env[63418]: _type = "Task" [ 1032.353963] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.362219] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245493, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.508472] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e56415-a817-05c5-3e69-fb8edf91e138, 'name': SearchDatastore_Task, 'duration_secs': 0.008778} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.509338] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ba920a-ece5-491b-af03-feb54c5b0917 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.515449] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1032.515449] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ab567b-f519-ee37-c9d6-18541acc3815" [ 1032.515449] env[63418]: _type = "Task" [ 1032.515449] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.524229] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ab567b-f519-ee37-c9d6-18541acc3815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.597098] env[63418]: DEBUG nova.objects.instance [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.709833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "5673b437-e993-47f8-8eaf-4be99869bfc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.710080] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.773651] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.773902] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.827373] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f867a6-e853-480a-92e6-d1cdb1d511b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.835618] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16287619-24f3-49f5-9ac8-000eab023943 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.869164] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecf53f2-59b3-42e8-8318-0ac2dc20562a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.878402] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245493, 'name': Rename_Task, 'duration_secs': 0.316348} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.879548] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b65a0ad-3fa1-4b30-8495-722e0df907b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.883038] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.883263] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-932a8a93-e1fe-481e-8b53-218e3d4828a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.894544] env[63418]: DEBUG nova.compute.provider_tree [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.896613] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1032.896613] env[63418]: value = "task-1245494" [ 1032.896613] env[63418]: _type = "Task" [ 1032.896613] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.903955] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245494, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.026354] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52ab567b-f519-ee37-c9d6-18541acc3815, 'name': SearchDatastore_Task, 'duration_secs': 0.011626} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.026638] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.026909] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 2dae6f5b-9bdc-4862-8654-2684ea22c6ac/2dae6f5b-9bdc-4862-8654-2684ea22c6ac.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.027210] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5aa2bf2d-d749-4088-be77-aeae4c2cdaad {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.035718] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1033.035718] env[63418]: value = "task-1245495" [ 1033.035718] env[63418]: _type = "Task" [ 1033.035718] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.045186] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245495, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.101884] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.102077] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.102254] env[63418]: DEBUG nova.network.neutron [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.102435] env[63418]: DEBUG nova.objects.instance [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'info_cache' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.214282] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1033.279266] env[63418]: DEBUG nova.compute.utils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.399302] env[63418]: DEBUG nova.scheduler.client.report [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1033.420304] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245494, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.548301] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245495, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.605610] env[63418]: DEBUG nova.objects.base [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1033.735299] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.784505] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.911179] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.215s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.916384] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.181s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.917848] env[63418]: INFO nova.compute.claims [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.920228] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245494, 'name': PowerOnVM_Task, 'duration_secs': 0.575407} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.920965] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.921060] env[63418]: INFO nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1033.921198] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1033.922373] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b13839d-e09a-4277-9b33-230cbe01f851 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.949026] env[63418]: INFO nova.network.neutron [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1034.046267] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245495, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514266} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.046523] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 2dae6f5b-9bdc-4862-8654-2684ea22c6ac/2dae6f5b-9bdc-4862-8654-2684ea22c6ac.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1034.046737] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.046981] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-405a9819-faf9-47c4-a65f-678196100f81 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.053318] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1034.053318] env[63418]: value = "task-1245497" [ 1034.053318] env[63418]: _type = "Task" [ 1034.053318] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.060975] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245497, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.364263] env[63418]: DEBUG nova.network.neutron [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [{"id": "01df43f5-4069-46a0-b1e6-4baff131e04e", "address": "fa:16:3e:69:39:54", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01df43f5-40", "ovs_interfaceid": "01df43f5-4069-46a0-b1e6-4baff131e04e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.440936] env[63418]: INFO nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Took 12.13 seconds to build instance. [ 1034.564614] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245497, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064932} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.564973] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.566129] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a4492a-eeec-424e-bc92-ba651706078a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.590463] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 2dae6f5b-9bdc-4862-8654-2684ea22c6ac/2dae6f5b-9bdc-4862-8654-2684ea22c6ac.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.591157] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5a35bfc-1630-49cb-8dac-0a139ebeb3e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.611375] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1034.611375] env[63418]: value = "task-1245498" [ 1034.611375] env[63418]: _type = "Task" [ 1034.611375] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.619757] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245498, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.851752] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.851987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.852259] env[63418]: INFO nova.compute.manager [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Attaching volume 6e5bcd2d-214c-4a04-8d2a-28ee190996b5 to /dev/sdb [ 1034.866344] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "refresh_cache-cbcbaf4f-ee24-4072-83ae-ffde59478928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.886638] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8958193a-c26c-4fc5-90d2-1c888c769ae0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.897424] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017a89ac-a6b7-4ca2-8b4c-74c795a8eb25 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.913247] env[63418]: DEBUG nova.virt.block_device [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating existing volume attachment record: 955c28b7-1df4-4ed1-b69e-1e4a2e7ea787 {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1034.943214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.643s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.066789] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ddf925-5712-40ee-88cb-17d4bb7cfb03 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.075234] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e19895-daa6-4aa8-95f2-5155216d7b6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.107358] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517999ce-646e-4e6b-880c-b921297032d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.118678] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c2503d-c420-4c49-9528-e460640aa2fd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.125918] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245498, 'name': ReconfigVM_Task, 'duration_secs': 0.28544} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.126587] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 2dae6f5b-9bdc-4862-8654-2684ea22c6ac/2dae6f5b-9bdc-4862-8654-2684ea22c6ac.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.127225] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4d35b60-2f0e-42a3-92a7-718f9fdbe6e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.136669] env[63418]: DEBUG nova.compute.provider_tree [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.143673] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1035.143673] env[63418]: value = "task-1245500" [ 1035.143673] env[63418]: _type = "Task" [ 1035.143673] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.151612] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245500, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.361879] env[63418]: DEBUG nova.compute.manager [req-468b56a2-8ae3-4499-9ddb-66698486a23a req-a4885a2b-97f4-4dc7-94e6-0b84471e01ca service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-vif-plugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1035.362088] env[63418]: DEBUG oslo_concurrency.lockutils [req-468b56a2-8ae3-4499-9ddb-66698486a23a req-a4885a2b-97f4-4dc7-94e6-0b84471e01ca service nova] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.362294] env[63418]: DEBUG oslo_concurrency.lockutils [req-468b56a2-8ae3-4499-9ddb-66698486a23a req-a4885a2b-97f4-4dc7-94e6-0b84471e01ca service nova] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.362498] env[63418]: DEBUG oslo_concurrency.lockutils [req-468b56a2-8ae3-4499-9ddb-66698486a23a req-a4885a2b-97f4-4dc7-94e6-0b84471e01ca service nova] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.362692] env[63418]: DEBUG nova.compute.manager [req-468b56a2-8ae3-4499-9ddb-66698486a23a req-a4885a2b-97f4-4dc7-94e6-0b84471e01ca service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] No waiting events found dispatching network-vif-plugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1035.363054] env[63418]: WARNING nova.compute.manager [req-468b56a2-8ae3-4499-9ddb-66698486a23a req-a4885a2b-97f4-4dc7-94e6-0b84471e01ca service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received unexpected event network-vif-plugged-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 for instance with vm_state shelved_offloaded and task_state spawning. [ 1035.463539] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.463929] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.463929] env[63418]: DEBUG nova.network.neutron [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.564302] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1035.564598] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268539', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'name': 'volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98f51fcd-951c-4c16-bc20-efc62ef359f5', 'attached_at': '', 'detached_at': '', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'serial': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1035.566017] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127dd6fc-9ff2-471d-a744-aff6c01b466b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.593427] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01640394-5394-4e1f-8c8c-73746544ea7b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.620612] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6/volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.620963] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5ca01da-182d-4798-ae7c-01541cbb6357 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.639133] env[63418]: DEBUG nova.scheduler.client.report [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1035.644712] env[63418]: DEBUG oslo_vmware.api [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1035.644712] env[63418]: value = "task-1245503" [ 1035.644712] env[63418]: _type = "Task" [ 1035.644712] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.662516] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245500, 'name': Rename_Task, 'duration_secs': 0.157408} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.662762] env[63418]: DEBUG oslo_vmware.api [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245503, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.663032] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.663293] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bd09219-73fd-4ec8-83a6-82791a4c97be {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.671435] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1035.671435] env[63418]: value = "task-1245504" [ 1035.671435] env[63418]: _type = "Task" [ 1035.671435] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.679744] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.873033] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.873394] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1018bd1a-74ef-4b00-9fba-5aa731a514fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.883029] env[63418]: DEBUG oslo_vmware.api [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1035.883029] env[63418]: value = "task-1245505" [ 1035.883029] env[63418]: _type = "Task" [ 1035.883029] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.893377] env[63418]: DEBUG oslo_vmware.api [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.146639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.147361] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1036.160538] env[63418]: DEBUG oslo_vmware.api [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245503, 'name': ReconfigVM_Task, 'duration_secs': 0.423931} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.160895] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Reconfigured VM instance instance-00000061 to attach disk [datastore2] volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6/volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.166482] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15c1d984-0c4b-4ded-83c6-ab8cd2b8cfc2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.186685] env[63418]: DEBUG oslo_vmware.api [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1036.186685] env[63418]: value = "task-1245506" [ 1036.186685] env[63418]: _type = "Task" [ 1036.186685] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.190136] env[63418]: DEBUG oslo_vmware.api [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245504, 'name': PowerOnVM_Task, 'duration_secs': 0.504158} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.193157] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.193374] env[63418]: INFO nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Took 7.41 seconds to spawn the instance on the hypervisor. [ 1036.193631] env[63418]: DEBUG nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1036.194932] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2617f4-85c9-4553-868b-3d81bc452d9e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.207318] env[63418]: DEBUG oslo_vmware.api [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245506, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.219753] env[63418]: DEBUG nova.network.neutron [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.394788] env[63418]: DEBUG oslo_vmware.api [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245505, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.655516] env[63418]: DEBUG nova.compute.utils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.656962] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1036.657948] env[63418]: DEBUG nova.network.neutron [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1036.695507] env[63418]: DEBUG nova.policy [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.702912] env[63418]: DEBUG oslo_vmware.api [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245506, 'name': ReconfigVM_Task, 'duration_secs': 0.153405} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.703232] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268539', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'name': 'volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98f51fcd-951c-4c16-bc20-efc62ef359f5', 'attached_at': '', 'detached_at': '', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'serial': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1036.715970] env[63418]: INFO nova.compute.manager [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Took 14.38 seconds to build instance. [ 1036.722593] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.755795] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='dbf381d9db24119ea19bb0d5336129b6',container_format='bare',created_at=2024-10-10T13:46:06Z,direct_url=,disk_format='vmdk',id=8691809c-e40f-490b-bcba-aaf810e4f073,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1338330597-shelved',owner='0e9f537407b84d50a49600de59e72c86',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-10T13:46:20Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.756119] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.756293] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.756482] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.756633] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.756782] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.756993] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.757175] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.757350] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.757514] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.757690] env[63418]: DEBUG nova.virt.hardware [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.758567] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96122384-fb37-4112-95a8-85977c877b16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.767963] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9edd4f-2db7-4571-8b7e-96115438f90b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.783193] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:76:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cc10d6a-03cf-43e3-84f7-9160f8626ad4', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.790910] env[63418]: DEBUG oslo.service.loopingcall [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.791160] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1036.791370] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0ca5855-100e-4f30-be87-1eef529e882a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.811658] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.811658] env[63418]: value = "task-1245507" [ 1036.811658] env[63418]: _type = "Task" [ 1036.811658] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.819349] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245507, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.895594] env[63418]: DEBUG oslo_vmware.api [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245505, 'name': PowerOnVM_Task, 'duration_secs': 0.558978} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.895832] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.896046] env[63418]: DEBUG nova.compute.manager [None req-f7375412-99f6-45af-8786-de9217e8aae6 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1036.896987] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0fe7d5-2a49-4eb4-8f86-72ac18011d3e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.951998] env[63418]: DEBUG nova.network.neutron [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Successfully created port: 33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.162825] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1037.216540] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a805561a-e79b-4e2c-8f37-137afc381559 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.887s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.323500] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245507, 'name': CreateVM_Task, 'duration_secs': 0.396148} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.323735] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.324555] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.324767] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.325261] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.325607] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b829b59-fab4-49eb-a9ac-fb6dad3e7b80 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.330861] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1037.330861] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e0ce3d-d59f-a42f-bb82-8b9a688d2d64" [ 1037.330861] env[63418]: _type = "Task" [ 1037.330861] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.339222] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e0ce3d-d59f-a42f-bb82-8b9a688d2d64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.390445] env[63418]: DEBUG nova.compute.manager [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1037.390664] env[63418]: DEBUG nova.compute.manager [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing instance network info cache due to event network-changed-9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1037.391070] env[63418]: DEBUG oslo_concurrency.lockutils [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] Acquiring lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.391134] env[63418]: DEBUG oslo_concurrency.lockutils [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] Acquired lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.391310] env[63418]: DEBUG nova.network.neutron [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Refreshing network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.745158] env[63418]: DEBUG nova.objects.instance [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'flavor' on Instance uuid 98f51fcd-951c-4c16-bc20-efc62ef359f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.842330] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.843212] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Processing image 8691809c-e40f-490b-bcba-aaf810e4f073 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.843212] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.843212] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.843212] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.843547] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3144ed56-5d96-460b-b787-9efd091c36d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.852582] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.852899] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1037.853674] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-054d8f1f-f455-48c1-a4ca-4918babcca79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.859227] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1037.859227] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d6652f-c518-930f-55ee-fe268e4ec325" [ 1037.859227] env[63418]: _type = "Task" [ 1037.859227] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.868341] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52d6652f-c518-930f-55ee-fe268e4ec325, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.042286] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "d0633a7d-33b4-4613-9501-81cef2ab3680" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.042548] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.042763] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "d0633a7d-33b4-4613-9501-81cef2ab3680-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.042986] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.043181] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.047269] env[63418]: INFO nova.compute.manager [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Terminating instance [ 1038.111153] env[63418]: DEBUG nova.network.neutron [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updated VIF entry in instance network info cache for port 9cc10d6a-03cf-43e3-84f7-9160f8626ad4. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.111736] env[63418]: DEBUG nova.network.neutron [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [{"id": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "address": "fa:16:3e:c2:76:43", "network": {"id": "56e2197e-aae0-408f-9802-76e191f3bb05", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1512870750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e9f537407b84d50a49600de59e72c86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc10d6a-03", "ovs_interfaceid": "9cc10d6a-03cf-43e3-84f7-9160f8626ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.126867] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.127128] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.127354] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.127529] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.127699] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.130061] env[63418]: INFO nova.compute.manager [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Terminating instance [ 1038.170860] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1038.195552] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.195864] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.195949] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.196142] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.196296] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.196543] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.196662] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.196820] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.196990] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.197185] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.197362] env[63418]: DEBUG nova.virt.hardware [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.198253] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e743c0-06b0-4680-9ef3-a6f0010c67b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.206648] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c244f85-b311-487a-96a6-294138437d4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.249331] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1c6857a2-163e-4667-b4a4-55d8e349959a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.306s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.373729] env[63418]: DEBUG nova.compute.manager [req-6d440535-373a-4f83-a557-9b0c5428f242 req-07425a85-a03f-423f-847f-58e96f5569b9 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Received event network-vif-plugged-33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1038.375049] env[63418]: DEBUG oslo_concurrency.lockutils [req-6d440535-373a-4f83-a557-9b0c5428f242 req-07425a85-a03f-423f-847f-58e96f5569b9 service nova] Acquiring lock "5673b437-e993-47f8-8eaf-4be99869bfc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.375346] env[63418]: DEBUG oslo_concurrency.lockutils [req-6d440535-373a-4f83-a557-9b0c5428f242 req-07425a85-a03f-423f-847f-58e96f5569b9 service nova] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.375528] env[63418]: DEBUG oslo_concurrency.lockutils [req-6d440535-373a-4f83-a557-9b0c5428f242 req-07425a85-a03f-423f-847f-58e96f5569b9 service nova] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.375704] env[63418]: DEBUG nova.compute.manager [req-6d440535-373a-4f83-a557-9b0c5428f242 req-07425a85-a03f-423f-847f-58e96f5569b9 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] No waiting events found dispatching network-vif-plugged-33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1038.377022] env[63418]: WARNING nova.compute.manager [req-6d440535-373a-4f83-a557-9b0c5428f242 req-07425a85-a03f-423f-847f-58e96f5569b9 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Received unexpected event network-vif-plugged-33423185-af4d-47b4-bc7f-f09a7eee9d17 for instance with vm_state building and task_state spawning. [ 1038.381884] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1038.382128] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Fetch image to [datastore1] OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452/OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1038.382546] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Downloading stream optimized image 8691809c-e40f-490b-bcba-aaf810e4f073 to [datastore1] OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452/OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452.vmdk on the data store datastore1 as vApp {{(pid=63418) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1038.382742] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Downloading image file data 8691809c-e40f-490b-bcba-aaf810e4f073 to the ESX as VM named 'OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452' {{(pid=63418) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1038.466545] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1038.466545] env[63418]: value = "resgroup-9" [ 1038.466545] env[63418]: _type = "ResourcePool" [ 1038.466545] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1038.466868] env[63418]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-73f1e243-2a12-47eb-b8fc-1f3539a28c62 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.490700] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease: (returnval){ [ 1038.490700] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292a29-e9f8-2ebc-d73f-75462957c898" [ 1038.490700] env[63418]: _type = "HttpNfcLease" [ 1038.490700] env[63418]: } obtained for vApp import into resource pool (val){ [ 1038.490700] env[63418]: value = "resgroup-9" [ 1038.490700] env[63418]: _type = "ResourcePool" [ 1038.490700] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1038.490990] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the lease: (returnval){ [ 1038.490990] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292a29-e9f8-2ebc-d73f-75462957c898" [ 1038.490990] env[63418]: _type = "HttpNfcLease" [ 1038.490990] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1038.497692] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1038.497692] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292a29-e9f8-2ebc-d73f-75462957c898" [ 1038.497692] env[63418]: _type = "HttpNfcLease" [ 1038.497692] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1038.550975] env[63418]: DEBUG nova.compute.manager [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1038.551316] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.552334] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a17ba2b-e663-4772-baab-b031444ae284 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.562648] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.563016] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82727591-515d-4bba-a968-a9e4a65cfa51 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.573152] env[63418]: DEBUG oslo_vmware.api [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1038.573152] env[63418]: value = "task-1245510" [ 1038.573152] env[63418]: _type = "Task" [ 1038.573152] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.585471] env[63418]: DEBUG oslo_vmware.api [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.614798] env[63418]: DEBUG oslo_concurrency.lockutils [req-ab2d197b-a1ed-484d-8de5-3399a046ca15 req-c11be026-9644-4082-8fa7-596e265e18a9 service nova] Releasing lock "refresh_cache-d52db46b-2461-4bd0-be57-d414250aac7e" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.634290] env[63418]: DEBUG nova.compute.manager [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1038.634621] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.635946] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6ef023-7efe-4250-af84-d9ff9214ab68 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.644177] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.644460] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09ed00b6-e129-4c1e-b2dc-faec1502a291 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.651287] env[63418]: DEBUG oslo_vmware.api [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1038.651287] env[63418]: value = "task-1245511" [ 1038.651287] env[63418]: _type = "Task" [ 1038.651287] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.660803] env[63418]: DEBUG oslo_vmware.api [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.956278] env[63418]: DEBUG nova.network.neutron [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Successfully updated port: 33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.999394] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1038.999394] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292a29-e9f8-2ebc-d73f-75462957c898" [ 1038.999394] env[63418]: _type = "HttpNfcLease" [ 1038.999394] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1038.999698] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1038.999698] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52292a29-e9f8-2ebc-d73f-75462957c898" [ 1038.999698] env[63418]: _type = "HttpNfcLease" [ 1038.999698] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1039.000951] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfa72cd-4e5b-486f-b497-420c00a382cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.007809] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ded21-f9aa-f774-66ab-803ef0a8a00d/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1039.008024] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ded21-f9aa-f774-66ab-803ef0a8a00d/disk-0.vmdk. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1039.071966] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-480d9248-bd15-408e-bfa9-8f86332fece3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.085413] env[63418]: DEBUG oslo_vmware.api [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245510, 'name': PowerOffVM_Task, 'duration_secs': 0.285194} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.085725] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1039.085921] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1039.086203] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43e7251a-2538-4e9a-89e7-25a7f6169657 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.149799] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1039.150084] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1039.150317] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleting the datastore file [datastore2] d0633a7d-33b4-4613-9501-81cef2ab3680 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1039.150608] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96a03304-b61e-4907-98ee-96c319165f52 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.161336] env[63418]: DEBUG oslo_vmware.api [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245511, 'name': PowerOffVM_Task, 'duration_secs': 0.220777} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.162517] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1039.162701] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1039.163045] env[63418]: DEBUG oslo_vmware.api [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1039.163045] env[63418]: value = "task-1245513" [ 1039.163045] env[63418]: _type = "Task" [ 1039.163045] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.163270] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a3f505b-e6ad-425a-b837-8ad6bebfa523 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.172852] env[63418]: DEBUG oslo_vmware.api [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.229102] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1039.229342] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1039.229508] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleting the datastore file [datastore2] 2dae6f5b-9bdc-4862-8654-2684ea22c6ac {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1039.231267] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4584b8eb-fca9-4c32-bfae-e1c0e912cb8d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.237779] env[63418]: DEBUG oslo_vmware.api [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1039.237779] env[63418]: value = "task-1245515" [ 1039.237779] env[63418]: _type = "Task" [ 1039.237779] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.245784] env[63418]: DEBUG oslo_vmware.api [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.417238] env[63418]: DEBUG nova.compute.manager [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Received event network-changed-33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1039.417406] env[63418]: DEBUG nova.compute.manager [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Refreshing instance network info cache due to event network-changed-33423185-af4d-47b4-bc7f-f09a7eee9d17. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1039.418089] env[63418]: DEBUG oslo_concurrency.lockutils [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] Acquiring lock "refresh_cache-5673b437-e993-47f8-8eaf-4be99869bfc6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.418089] env[63418]: DEBUG oslo_concurrency.lockutils [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] Acquired lock "refresh_cache-5673b437-e993-47f8-8eaf-4be99869bfc6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.418089] env[63418]: DEBUG nova.network.neutron [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Refreshing network info cache for port 33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.460062] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-5673b437-e993-47f8-8eaf-4be99869bfc6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.462960] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1039.463206] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268541', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'name': 'volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a8b347c-e9ec-4355-bcf7-b7d205da194c', 'attached_at': '', 'detached_at': '', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'serial': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1039.464091] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad9a32e-0cdc-4559-a7c5-b25252cd1283 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.487816] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2469e3f-bce0-45e4-bd92-1804a8d68df7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.515643] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5/volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.515944] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad432981-f3cf-4eb8-bcfe-c88d53325e44 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.533775] env[63418]: DEBUG oslo_vmware.api [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1039.533775] env[63418]: value = "task-1245516" [ 1039.533775] env[63418]: _type = "Task" [ 1039.533775] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.541549] env[63418]: DEBUG oslo_vmware.api [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245516, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.678165] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "fe38702e-953d-4748-a523-401e798b2271" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.678411] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "fe38702e-953d-4748-a523-401e798b2271" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.679540] env[63418]: DEBUG oslo_vmware.api [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170717} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.681585] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.681823] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.682043] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.682230] env[63418]: INFO nova.compute.manager [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1039.682466] env[63418]: DEBUG oslo.service.loopingcall [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.682918] env[63418]: DEBUG nova.compute.manager [-] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1039.683010] env[63418]: DEBUG nova.network.neutron [-] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.750638] env[63418]: DEBUG oslo_vmware.api [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249208} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.750959] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.751172] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.751869] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.751869] env[63418]: INFO nova.compute.manager [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1039.751869] env[63418]: DEBUG oslo.service.loopingcall [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.753220] env[63418]: DEBUG nova.compute.manager [-] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1039.753311] env[63418]: DEBUG nova.network.neutron [-] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.959908] env[63418]: DEBUG nova.network.neutron [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1040.047040] env[63418]: DEBUG oslo_vmware.api [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245516, 'name': ReconfigVM_Task, 'duration_secs': 0.429624} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.048782] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to attach disk [datastore1] volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5/volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.055872] env[63418]: DEBUG nova.network.neutron [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.057399] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61274229-ccc6-4dcc-9561-502508506ee2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.072754] env[63418]: DEBUG oslo_concurrency.lockutils [req-0fce8eb7-3b30-425f-9722-94b0041c236d req-8bd859ef-416c-4852-b62b-d88480b8f648 service nova] Releasing lock "refresh_cache-5673b437-e993-47f8-8eaf-4be99869bfc6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.073508] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-5673b437-e993-47f8-8eaf-4be99869bfc6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.073680] env[63418]: DEBUG nova.network.neutron [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.079877] env[63418]: DEBUG oslo_vmware.api [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1040.079877] env[63418]: value = "task-1245517" [ 1040.079877] env[63418]: _type = "Task" [ 1040.079877] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.086648] env[63418]: DEBUG nova.compute.manager [req-136ef1d7-ff91-4efe-988e-aaf898eec479 req-f7c00c6c-2898-47c3-94f5-154a2840b2fa service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Received event network-vif-deleted-a5c86e79-a127-4e43-9d97-8c69ea66352a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1040.086869] env[63418]: INFO nova.compute.manager [req-136ef1d7-ff91-4efe-988e-aaf898eec479 req-f7c00c6c-2898-47c3-94f5-154a2840b2fa service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Neutron deleted interface a5c86e79-a127-4e43-9d97-8c69ea66352a; detaching it from the instance and deleting it from the info cache [ 1040.087075] env[63418]: DEBUG nova.network.neutron [req-136ef1d7-ff91-4efe-988e-aaf898eec479 req-f7c00c6c-2898-47c3-94f5-154a2840b2fa service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.094528] env[63418]: DEBUG oslo_vmware.api [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245517, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.180622] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1040.215895] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1040.216246] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ded21-f9aa-f774-66ab-803ef0a8a00d/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1040.217251] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6d68ab-8ae7-45f5-9604-92799c917b55 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.226657] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ded21-f9aa-f774-66ab-803ef0a8a00d/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1040.226657] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ded21-f9aa-f774-66ab-803ef0a8a00d/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1040.226920] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-621d48fc-a1e4-46a2-8d64-b1c0a8a8054b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.400364] env[63418]: DEBUG oslo_vmware.rw_handles [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ded21-f9aa-f774-66ab-803ef0a8a00d/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1040.400597] env[63418]: INFO nova.virt.vmwareapi.images [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Downloaded image file data 8691809c-e40f-490b-bcba-aaf810e4f073 [ 1040.401660] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef16ff06-55b3-432f-a12b-2dc6c0d3aa4e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.418581] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ee43dad-2ded-4f4a-8673-2f10c6c44aee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.446139] env[63418]: INFO nova.virt.vmwareapi.images [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] The imported VM was unregistered [ 1040.448557] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1040.448784] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Creating directory with path [datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.449052] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75e0daa3-f2e3-4e7d-9a5a-08ffd41dee53 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.470298] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Created directory with path [datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073 {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.470693] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452/OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452.vmdk to [datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk. {{(pid=63418) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1040.470985] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-beb876a1-cbd9-49e4-9848-730e4706b1a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.478390] env[63418]: DEBUG nova.network.neutron [-] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.479730] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1040.479730] env[63418]: value = "task-1245519" [ 1040.479730] env[63418]: _type = "Task" [ 1040.479730] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.489316] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.576280] env[63418]: DEBUG nova.network.neutron [-] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.593021] env[63418]: DEBUG oslo_vmware.api [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245517, 'name': ReconfigVM_Task, 'duration_secs': 0.17953} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.593307] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1479d5d0-b571-46fc-8591-a142b48d6045 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.595191] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268541', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'name': 'volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a8b347c-e9ec-4355-bcf7-b7d205da194c', 'attached_at': '', 'detached_at': '', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'serial': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1040.603425] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca4ed36-9c97-4322-9a5f-6c057e8c90f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.616102] env[63418]: DEBUG nova.network.neutron [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1040.634594] env[63418]: DEBUG nova.compute.manager [req-136ef1d7-ff91-4efe-988e-aaf898eec479 req-f7c00c6c-2898-47c3-94f5-154a2840b2fa service nova] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Detach interface failed, port_id=a5c86e79-a127-4e43-9d97-8c69ea66352a, reason: Instance 2dae6f5b-9bdc-4862-8654-2684ea22c6ac could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1040.704799] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.705161] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.706720] env[63418]: INFO nova.compute.claims [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.768362] env[63418]: DEBUG nova.network.neutron [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Updating instance_info_cache with network_info: [{"id": "33423185-af4d-47b4-bc7f-f09a7eee9d17", "address": "fa:16:3e:4f:9d:e1", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33423185-af", "ovs_interfaceid": "33423185-af4d-47b4-bc7f-f09a7eee9d17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.981856] env[63418]: INFO nova.compute.manager [-] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Took 1.30 seconds to deallocate network for instance. [ 1040.996020] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.081264] env[63418]: INFO nova.compute.manager [-] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Took 1.33 seconds to deallocate network for instance. [ 1041.271403] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-5673b437-e993-47f8-8eaf-4be99869bfc6" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.271734] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Instance network_info: |[{"id": "33423185-af4d-47b4-bc7f-f09a7eee9d17", "address": "fa:16:3e:4f:9d:e1", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33423185-af", "ovs_interfaceid": "33423185-af4d-47b4-bc7f-f09a7eee9d17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1041.272199] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:9d:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33423185-af4d-47b4-bc7f-f09a7eee9d17', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.279767] env[63418]: DEBUG oslo.service.loopingcall [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.279975] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1041.280221] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b748f3c-2687-48c8-b2bb-991393e9ad75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.299928] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.299928] env[63418]: value = "task-1245520" [ 1041.299928] env[63418]: _type = "Task" [ 1041.299928] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.309142] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245520, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.447055] env[63418]: DEBUG nova.compute.manager [req-3be03d0a-ddf1-4bb0-9bb3-260f87080728 req-c81d22dd-10d3-477f-8b53-f6be69029976 service nova] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Received event network-vif-deleted-a5d70b5d-10c4-4709-8fe1-f84cf0e84489 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1041.493305] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.493613] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.588489] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.631066] env[63418]: DEBUG nova.objects.instance [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'flavor' on Instance uuid 5a8b347c-e9ec-4355-bcf7-b7d205da194c {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.812587] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245520, 'name': CreateVM_Task, 'duration_secs': 0.363189} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.812804] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1041.813592] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.813783] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.814144] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1041.814434] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e0f4ced-f307-4b04-95e3-1f45063225cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.819762] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1041.819762] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5277f6d5-6f3d-bf7b-8e05-dc4174b6fe0d" [ 1041.819762] env[63418]: _type = "Task" [ 1041.819762] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.831869] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5277f6d5-6f3d-bf7b-8e05-dc4174b6fe0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.860423] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4818800d-c4fb-414f-9b91-175fcf8d56a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.869803] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdcd6b4-ca47-4243-83ae-26025bc235f7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.903247] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d28443b-7d6f-46cc-b06d-fd3107088a6d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.912017] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5076728-f3e2-410b-8c4f-71643810732f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.927303] env[63418]: DEBUG nova.compute.provider_tree [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.996089] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.137206] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e065869e-4d33-40ad-bdff-8dcabee8932c tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.285s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.335204] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5277f6d5-6f3d-bf7b-8e05-dc4174b6fe0d, 'name': SearchDatastore_Task, 'duration_secs': 0.084165} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.335639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.335919] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.336254] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.336438] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.336704] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.337052] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55bcb710-fa10-4648-96c0-28db980b709b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.353659] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.353888] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1042.354703] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d92c6759-7357-4bfe-9e3a-e9ce8340dd2f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.360817] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1042.360817] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f1baef-71db-a6eb-ca85-9f93baada047" [ 1042.360817] env[63418]: _type = "Task" [ 1042.360817] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.368917] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f1baef-71db-a6eb-ca85-9f93baada047, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.430688] env[63418]: DEBUG nova.scheduler.client.report [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1042.493452] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.871464] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52f1baef-71db-a6eb-ca85-9f93baada047, 'name': SearchDatastore_Task, 'duration_secs': 0.085112} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.872379] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eae799b9-6c87-4f83-9908-315b5fb90e91 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.877905] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1042.877905] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522865ec-3890-52a7-8af4-fc2dfa039788" [ 1042.877905] env[63418]: _type = "Task" [ 1042.877905] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.886124] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522865ec-3890-52a7-8af4-fc2dfa039788, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.935541] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.936146] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1042.940076] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.446s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.940076] env[63418]: DEBUG nova.objects.instance [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lazy-loading 'resources' on Instance uuid d0633a7d-33b4-4613-9501-81cef2ab3680 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.994368] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.164311] env[63418]: DEBUG nova.compute.manager [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Stashing vm_state: active {{(pid=63418) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1043.388554] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522865ec-3890-52a7-8af4-fc2dfa039788, 'name': SearchDatastore_Task, 'duration_secs': 0.078155} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.388787] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.389053] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5673b437-e993-47f8-8eaf-4be99869bfc6/5673b437-e993-47f8-8eaf-4be99869bfc6.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1043.389332] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0f66604-4da9-47a6-8349-883ac75b8d2a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.395636] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1043.395636] env[63418]: value = "task-1245521" [ 1043.395636] env[63418]: _type = "Task" [ 1043.395636] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.405315] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245521, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.443154] env[63418]: DEBUG nova.compute.utils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.447897] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1043.448109] env[63418]: DEBUG nova.network.neutron [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.491030] env[63418]: DEBUG nova.policy [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b29941866349482fb9e53dcf87cb1845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05ef1c6c74574217817c6ab14a022b91', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1043.497962] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.607020] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98109b9e-60fa-4856-9131-841cd0d442f5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.614242] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71304af4-f2c6-4f88-8430-bf334566585a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.645705] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06dac07-a208-4d60-a55b-2ea71f159776 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.653962] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8be7d95-2d05-49a1-9904-b93690dbe4ed {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.667634] env[63418]: DEBUG nova.compute.provider_tree [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1043.683957] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.907736] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245521, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.948307] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1043.957264] env[63418]: DEBUG nova.network.neutron [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Successfully created port: 9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.997862] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245519, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.147895} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.999033] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452/OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452.vmdk to [datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk. [ 1043.999248] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Cleaning up location [datastore1] OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1043.999490] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_9b295c7b-de37-4c6c-9e40-ca73534a4452 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.999796] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4f99dca-6b55-4c19-8682-4b1768abaf3f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.007958] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1044.007958] env[63418]: value = "task-1245522" [ 1044.007958] env[63418]: _type = "Task" [ 1044.007958] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.019395] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.191403] env[63418]: ERROR nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [req-7d394db2-050b-4e23-8f9d-4ab403178e80] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7d394db2-050b-4e23-8f9d-4ab403178e80"}]} [ 1044.207083] env[63418]: DEBUG nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1044.221249] env[63418]: DEBUG nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1044.221486] env[63418]: DEBUG nova.compute.provider_tree [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1044.231943] env[63418]: DEBUG nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1044.248686] env[63418]: DEBUG nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1044.387084] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad46598-5976-495a-98a3-d22c70a20a27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.394882] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d111554b-13b0-44d3-ac3e-a5f110a8e231 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.406259] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245521, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.819468} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.430011] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 5673b437-e993-47f8-8eaf-4be99869bfc6/5673b437-e993-47f8-8eaf-4be99869bfc6.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.430281] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.430762] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fcc41bd3-3058-42e2-978a-be16d31202f7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.433536] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2309bd5-2e1d-4b20-85e0-c1f10b6ba392 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.441372] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8629f2-4495-44cf-a041-032e621ab37d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.445332] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1044.445332] env[63418]: value = "task-1245523" [ 1044.445332] env[63418]: _type = "Task" [ 1044.445332] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.459796] env[63418]: DEBUG nova.compute.provider_tree [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1044.465423] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245523, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.518399] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191302} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.519066] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.519066] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.519066] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk to [datastore1] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.519375] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d8cc8bb-73f1-461b-a9bf-bcad40cf2435 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.525897] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1044.525897] env[63418]: value = "task-1245524" [ 1044.525897] env[63418]: _type = "Task" [ 1044.525897] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.532748] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.960634] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245523, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064646} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.960921] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1044.962175] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1044.964821] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c2347f-4f4f-4fc6-b120-e9e3aceea7e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.995939] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 5673b437-e993-47f8-8eaf-4be99869bfc6/5673b437-e993-47f8-8eaf-4be99869bfc6.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.997037] env[63418]: DEBUG nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 139 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1044.997295] env[63418]: DEBUG nova.compute.provider_tree [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 139 to 140 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1044.997617] env[63418]: DEBUG nova.compute.provider_tree [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.002553] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a426b8a1-dd6c-49f3-b028-96d565a53a67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.023864] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.024177] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.024365] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.024656] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.024832] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.024984] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.025239] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.025411] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.025595] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.025758] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.025934] env[63418]: DEBUG nova.virt.hardware [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.027168] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eeff67d-6e1e-45df-8d90-4f43f4b022af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.034582] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1045.034582] env[63418]: value = "task-1245525" [ 1045.034582] env[63418]: _type = "Task" [ 1045.034582] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.041831] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc7f53b-b09e-4e99-80db-c3f2f6ffedd7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.052602] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245525, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.052862] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245524, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.378500] env[63418]: DEBUG nova.compute.manager [req-efc063a8-9d85-4457-be88-845ac07bb9d3 req-219ae342-3c16-46cb-aba7-9723c5169d86 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Received event network-vif-plugged-9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1045.378745] env[63418]: DEBUG oslo_concurrency.lockutils [req-efc063a8-9d85-4457-be88-845ac07bb9d3 req-219ae342-3c16-46cb-aba7-9723c5169d86 service nova] Acquiring lock "fe38702e-953d-4748-a523-401e798b2271-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.378938] env[63418]: DEBUG oslo_concurrency.lockutils [req-efc063a8-9d85-4457-be88-845ac07bb9d3 req-219ae342-3c16-46cb-aba7-9723c5169d86 service nova] Lock "fe38702e-953d-4748-a523-401e798b2271-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.382044] env[63418]: DEBUG oslo_concurrency.lockutils [req-efc063a8-9d85-4457-be88-845ac07bb9d3 req-219ae342-3c16-46cb-aba7-9723c5169d86 service nova] Lock "fe38702e-953d-4748-a523-401e798b2271-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.382318] env[63418]: DEBUG nova.compute.manager [req-efc063a8-9d85-4457-be88-845ac07bb9d3 req-219ae342-3c16-46cb-aba7-9723c5169d86 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] No waiting events found dispatching network-vif-plugged-9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1045.382514] env[63418]: WARNING nova.compute.manager [req-efc063a8-9d85-4457-be88-845ac07bb9d3 req-219ae342-3c16-46cb-aba7-9723c5169d86 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Received unexpected event network-vif-plugged-9e929283-d11e-4069-ab19-e7195f2cfd1b for instance with vm_state building and task_state spawning. [ 1045.518407] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.579s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.520871] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.933s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.521218] env[63418]: DEBUG nova.objects.instance [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lazy-loading 'resources' on Instance uuid 2dae6f5b-9bdc-4862-8654-2684ea22c6ac {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.537895] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245524, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.538994] env[63418]: INFO nova.scheduler.client.report [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted allocations for instance d0633a7d-33b4-4613-9501-81cef2ab3680 [ 1045.553840] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245525, 'name': ReconfigVM_Task, 'duration_secs': 0.329103} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.554631] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 5673b437-e993-47f8-8eaf-4be99869bfc6/5673b437-e993-47f8-8eaf-4be99869bfc6.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.555404] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c4e5bd0-b45b-4b63-9bf6-492bb1be1021 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.563123] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1045.563123] env[63418]: value = "task-1245526" [ 1045.563123] env[63418]: _type = "Task" [ 1045.563123] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.572355] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245526, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.840525] env[63418]: DEBUG nova.network.neutron [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Successfully updated port: 9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.038779] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245524, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.052427] env[63418]: DEBUG oslo_concurrency.lockutils [None req-96c78883-92ce-4c3c-a9ce-91682c51cc1a tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "d0633a7d-33b4-4613-9501-81cef2ab3680" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.010s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.074760] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245526, 'name': Rename_Task, 'duration_secs': 0.163071} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.075053] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.075350] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa6f9586-87f4-4f03-9809-b80be40585ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.082308] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1046.082308] env[63418]: value = "task-1245527" [ 1046.082308] env[63418]: _type = "Task" [ 1046.082308] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.093081] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.178536] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddec319-04d6-43a1-a8a5-d0a61dd49d94 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.187018] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b788b780-36db-4cad-ac2d-06e8109ef440 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.218175] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c49acb0-6888-4b9c-987c-5a73dda61173 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.225946] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad161e77-d6f4-43be-98c8-44878172d8c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.239494] env[63418]: DEBUG nova.compute.provider_tree [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.345832] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.345992] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.346154] env[63418]: DEBUG nova.network.neutron [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.540739] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245524, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.593266] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245527, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.742256] env[63418]: DEBUG nova.scheduler.client.report [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1046.910148] env[63418]: DEBUG nova.network.neutron [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.040186] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245524, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.294591} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.040476] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8691809c-e40f-490b-bcba-aaf810e4f073/8691809c-e40f-490b-bcba-aaf810e4f073.vmdk to [datastore1] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1047.041459] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6911aa5-b73e-40ce-96f0-1f9631e66b96 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.065202] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.065493] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66d7d67c-07f9-47f9-a9bd-3430d50da683 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.087566] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1047.087566] env[63418]: value = "task-1245528" [ 1047.087566] env[63418]: _type = "Task" [ 1047.087566] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.094041] env[63418]: DEBUG oslo_vmware.api [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245527, 'name': PowerOnVM_Task, 'duration_secs': 0.624609} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.094836] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1047.095118] env[63418]: INFO nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Took 8.92 seconds to spawn the instance on the hypervisor. [ 1047.095379] env[63418]: DEBUG nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1047.096141] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bee83d-3824-46a6-95b4-915687a01f0f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.101442] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245528, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.148433] env[63418]: DEBUG nova.network.neutron [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Updating instance_info_cache with network_info: [{"id": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "address": "fa:16:3e:4b:1f:84", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e929283-d1", "ovs_interfaceid": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.247367] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.726s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.249670] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.566s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.276088] env[63418]: INFO nova.scheduler.client.report [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted allocations for instance 2dae6f5b-9bdc-4862-8654-2684ea22c6ac [ 1047.599126] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.616023] env[63418]: DEBUG nova.compute.manager [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Received event network-changed-9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1047.616275] env[63418]: DEBUG nova.compute.manager [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Refreshing instance network info cache due to event network-changed-9e929283-d11e-4069-ab19-e7195f2cfd1b. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1047.616553] env[63418]: DEBUG oslo_concurrency.lockutils [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] Acquiring lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.618672] env[63418]: INFO nova.compute.manager [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Took 13.90 seconds to build instance. [ 1047.651089] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.651382] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Instance network_info: |[{"id": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "address": "fa:16:3e:4b:1f:84", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e929283-d1", "ovs_interfaceid": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1047.651904] env[63418]: DEBUG oslo_concurrency.lockutils [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] Acquired lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.652104] env[63418]: DEBUG nova.network.neutron [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Refreshing network info cache for port 9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.653216] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:1f:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'af454577-0e89-41a3-a9f2-f39716f62fd5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e929283-d11e-4069-ab19-e7195f2cfd1b', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.660734] env[63418]: DEBUG oslo.service.loopingcall [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.661534] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe38702e-953d-4748-a523-401e798b2271] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.661763] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-180b001c-b9e1-4c96-be99-04e0c8e564e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.682016] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.682016] env[63418]: value = "task-1245529" [ 1047.682016] env[63418]: _type = "Task" [ 1047.682016] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.691372] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245529, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.754584] env[63418]: INFO nova.compute.claims [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.783693] env[63418]: DEBUG oslo_concurrency.lockutils [None req-f29aac67-2990-4678-b913-ecfd9dae133b tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "2dae6f5b-9bdc-4862-8654-2684ea22c6ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.656s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.099363] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245528, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.121804] env[63418]: DEBUG oslo_concurrency.lockutils [None req-74f9d2c9-fead-4223-8898-6adaeb3b968d tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.411s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.192508] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245529, 'name': CreateVM_Task, 'duration_secs': 0.374122} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.192927] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe38702e-953d-4748-a523-401e798b2271] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.193744] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.194066] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.194512] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.195117] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99bbbff4-8946-4d40-a445-99e20ed624c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.202184] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1048.202184] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529ff444-ccb0-541a-3ed3-c0413736492d" [ 1048.202184] env[63418]: _type = "Task" [ 1048.202184] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.207772] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529ff444-ccb0-541a-3ed3-c0413736492d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.262208] env[63418]: INFO nova.compute.resource_tracker [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating resource usage from migration 1500a04c-32e3-42be-a525-4d03eac3038b [ 1048.378179] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267d0f7b-4979-41df-8019-5f9f85001796 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.385367] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f99bee8-f886-4353-a6d3-524ac3fc10df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.419366] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677a37ca-8cc3-4903-9c0d-6313996e0a08 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.428484] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f662e8b-5ddd-4fbd-9a1c-b96789d715d1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.443017] env[63418]: DEBUG nova.compute.provider_tree [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.551996] env[63418]: DEBUG nova.network.neutron [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Updated VIF entry in instance network info cache for port 9e929283-d11e-4069-ab19-e7195f2cfd1b. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.552394] env[63418]: DEBUG nova.network.neutron [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Updating instance_info_cache with network_info: [{"id": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "address": "fa:16:3e:4b:1f:84", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e929283-d1", "ovs_interfaceid": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.600094] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245528, 'name': ReconfigVM_Task, 'duration_secs': 1.047088} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.600413] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Reconfigured VM instance instance-0000005e to attach disk [datastore1] d52db46b-2461-4bd0-be57-d414250aac7e/d52db46b-2461-4bd0-be57-d414250aac7e.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.601311] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e0111bb-b479-4f8e-9103-35defeef6325 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.607908] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1048.607908] env[63418]: value = "task-1245530" [ 1048.607908] env[63418]: _type = "Task" [ 1048.607908] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.615335] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245530, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.713455] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529ff444-ccb0-541a-3ed3-c0413736492d, 'name': SearchDatastore_Task, 'duration_secs': 0.008539} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.713708] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.714397] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.714470] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.715186] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.715186] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.716457] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12a715c5-6815-4743-9177-1ef0161ecc3b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.725557] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.725795] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.726546] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a4f63f6-d63a-429a-ab17-faf19ff06b50 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.731457] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1048.731457] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5282ca27-ae06-9bc6-6bc3-1c0a37c9ac2a" [ 1048.731457] env[63418]: _type = "Task" [ 1048.731457] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.739406] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5282ca27-ae06-9bc6-6bc3-1c0a37c9ac2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.948298] env[63418]: DEBUG nova.scheduler.client.report [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1049.056240] env[63418]: DEBUG oslo_concurrency.lockutils [req-3302bdee-f0ea-4561-bbcc-702d5b578a2d req-16462c91-53bb-4515-b200-c429151d6282 service nova] Releasing lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.117533] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245530, 'name': Rename_Task, 'duration_secs': 0.175502} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.117810] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.118192] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d1f9228-052e-4a5b-97bd-bf0cbef6a769 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.124238] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1049.124238] env[63418]: value = "task-1245531" [ 1049.124238] env[63418]: _type = "Task" [ 1049.124238] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.131884] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245531, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.241954] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "960184bb-0e6f-44ab-a119-fb43c27df929" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.241954] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "960184bb-0e6f-44ab-a119-fb43c27df929" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.246893] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5282ca27-ae06-9bc6-6bc3-1c0a37c9ac2a, 'name': SearchDatastore_Task, 'duration_secs': 0.009594} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.247506] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-575c4372-e786-4e49-815a-f498a1452285 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.252847] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1049.252847] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c27e64-b097-9e99-3e55-3dc676bd71c9" [ 1049.252847] env[63418]: _type = "Task" [ 1049.252847] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.263132] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c27e64-b097-9e99-3e55-3dc676bd71c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.267340] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "1e6828a8-02ea-434f-b587-8d4c6854d944" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.267565] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.321872] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "5673b437-e993-47f8-8eaf-4be99869bfc6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.322087] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.322307] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "5673b437-e993-47f8-8eaf-4be99869bfc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.322497] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.322670] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.324911] env[63418]: INFO nova.compute.manager [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Terminating instance [ 1049.453939] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.204s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.454238] env[63418]: INFO nova.compute.manager [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Migrating [ 1049.634920] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245531, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.747803] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1049.763526] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c27e64-b097-9e99-3e55-3dc676bd71c9, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.763795] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.764068] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] fe38702e-953d-4748-a523-401e798b2271/fe38702e-953d-4748-a523-401e798b2271.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.764379] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7a26363-ef73-4965-b924-b8feb9d5f47e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.770150] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1049.774054] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1049.774054] env[63418]: value = "task-1245532" [ 1049.774054] env[63418]: _type = "Task" [ 1049.774054] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.782123] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.828589] env[63418]: DEBUG nova.compute.manager [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1049.828813] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.829723] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2d7cd3-a1a9-4fd2-9877-ab9409ba3092 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.837965] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.838257] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad110a52-391b-4f4e-858a-aeb5f672340c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.843970] env[63418]: DEBUG oslo_vmware.api [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1049.843970] env[63418]: value = "task-1245533" [ 1049.843970] env[63418]: _type = "Task" [ 1049.843970] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.851810] env[63418]: DEBUG oslo_vmware.api [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.971190] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.971190] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.971190] env[63418]: DEBUG nova.network.neutron [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1050.137465] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245531, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.274583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.274778] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.276815] env[63418]: INFO nova.compute.claims [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.294230] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245532, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.298930] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.353367] env[63418]: DEBUG oslo_vmware.api [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245533, 'name': PowerOffVM_Task, 'duration_secs': 0.171837} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.353662] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.353807] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.354070] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e506b0dd-1859-42fe-956b-20597cdda6b9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.439620] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.440689] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.440689] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore1] 5673b437-e993-47f8-8eaf-4be99869bfc6 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.440689] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d15caa60-ff21-4eab-819d-8c0be5fc40b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.447439] env[63418]: DEBUG oslo_vmware.api [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1050.447439] env[63418]: value = "task-1245535" [ 1050.447439] env[63418]: _type = "Task" [ 1050.447439] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.456232] env[63418]: DEBUG oslo_vmware.api [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245535, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.634888] env[63418]: DEBUG oslo_vmware.api [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245531, 'name': PowerOnVM_Task, 'duration_secs': 1.166666} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.635176] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.735359] env[63418]: DEBUG nova.compute.manager [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1050.736281] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cda2a7e-b58f-4200-8c3e-862991205159 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.754569] env[63418]: DEBUG nova.network.neutron [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.794756] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540588} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.795037] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] fe38702e-953d-4748-a523-401e798b2271/fe38702e-953d-4748-a523-401e798b2271.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1050.795316] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.795568] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-170cfc44-ed97-444a-aa8c-b2ae195ab660 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.802692] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1050.802692] env[63418]: value = "task-1245536" [ 1050.802692] env[63418]: _type = "Task" [ 1050.802692] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.809839] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245536, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.956585] env[63418]: DEBUG oslo_vmware.api [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245535, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149281} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.956890] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.957098] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.957286] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.957466] env[63418]: INFO nova.compute.manager [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1050.957714] env[63418]: DEBUG oslo.service.loopingcall [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.957902] env[63418]: DEBUG nova.compute.manager [-] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1050.957995] env[63418]: DEBUG nova.network.neutron [-] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.239853] env[63418]: DEBUG nova.compute.manager [req-981eb3e4-68a9-4a16-90cc-8c56f60a2ada req-1bf132b1-5b1d-4e45-87d3-6a26ed1846e5 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Received event network-vif-deleted-33423185-af4d-47b4-bc7f-f09a7eee9d17 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1051.240077] env[63418]: INFO nova.compute.manager [req-981eb3e4-68a9-4a16-90cc-8c56f60a2ada req-1bf132b1-5b1d-4e45-87d3-6a26ed1846e5 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Neutron deleted interface 33423185-af4d-47b4-bc7f-f09a7eee9d17; detaching it from the instance and deleting it from the info cache [ 1051.240300] env[63418]: DEBUG nova.network.neutron [req-981eb3e4-68a9-4a16-90cc-8c56f60a2ada req-1bf132b1-5b1d-4e45-87d3-6a26ed1846e5 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.253937] env[63418]: DEBUG oslo_concurrency.lockutils [None req-17c54c84-df28-47da-a76f-d6c835bd02a5 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.749s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.256639] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.313631] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245536, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.313039} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.313898] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.314667] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502e6b81-deb4-4baa-9529-56bb9bef79bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.338113] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] fe38702e-953d-4748-a523-401e798b2271/fe38702e-953d-4748-a523-401e798b2271.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.340656] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bac6158-5942-4081-9119-680cbdd0c3e1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.359280] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1051.359280] env[63418]: value = "task-1245537" [ 1051.359280] env[63418]: _type = "Task" [ 1051.359280] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.368073] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245537, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.452096] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcc7e4f-39f1-4c6c-85e0-7c81444e6afe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.459426] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdac67f-a70a-49af-afdf-fd4da0220cfd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.491809] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329ee279-df1a-4d39-9c1c-a129febd321c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.499534] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87d5dbd-62fa-431d-b555-3cb07fd0f785 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.513430] env[63418]: DEBUG nova.compute.provider_tree [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.714061] env[63418]: DEBUG nova.network.neutron [-] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.743311] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d38552f7-4ef8-4ea1-8127-c92c0d5905d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.752012] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485011d1-8ab1-4d39-9cb9-e53ae61d9ba8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.782908] env[63418]: DEBUG nova.compute.manager [req-981eb3e4-68a9-4a16-90cc-8c56f60a2ada req-1bf132b1-5b1d-4e45-87d3-6a26ed1846e5 service nova] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Detach interface failed, port_id=33423185-af4d-47b4-bc7f-f09a7eee9d17, reason: Instance 5673b437-e993-47f8-8eaf-4be99869bfc6 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1051.873691] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245537, 'name': ReconfigVM_Task, 'duration_secs': 0.255681} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.873691] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Reconfigured VM instance instance-0000006a to attach disk [datastore2] fe38702e-953d-4748-a523-401e798b2271/fe38702e-953d-4748-a523-401e798b2271.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.873691] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c480902-e2b2-4552-b8e5-ae841d7e7e27 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.879801] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1051.879801] env[63418]: value = "task-1245538" [ 1051.879801] env[63418]: _type = "Task" [ 1051.879801] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.889314] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245538, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.032649] env[63418]: ERROR nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [req-d5f19edf-6384-4e9d-9ab2-78f952d4cbab] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ac9de28-4c58-4fc2-8a3d-711092e3c63c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d5f19edf-6384-4e9d-9ab2-78f952d4cbab"}]} [ 1052.048318] env[63418]: DEBUG nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Refreshing inventories for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1052.061108] env[63418]: DEBUG nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating ProviderTree inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1052.061327] env[63418]: DEBUG nova.compute.provider_tree [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1052.071232] env[63418]: DEBUG nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Refreshing aggregate associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, aggregates: None {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1052.088234] env[63418]: DEBUG nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Refreshing trait associations for resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63418) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1052.206641] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07436bbf-dd5b-4369-b94c-e26f475e5514 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.214126] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec12698-5468-4322-ae77-4d5d5b79f87f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.217212] env[63418]: INFO nova.compute.manager [-] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Took 1.26 seconds to deallocate network for instance. [ 1052.246499] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dfe669-17e7-4073-b0a6-2d87a1415a9b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.254032] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ca09bd-c92d-4825-9933-d2d2157514aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.267100] env[63418]: DEBUG nova.compute.provider_tree [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1052.389682] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245538, 'name': Rename_Task, 'duration_secs': 0.126002} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.389945] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1052.390205] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37b9e405-955f-4fc8-a3a5-32f17c0d11b5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.395591] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1052.395591] env[63418]: value = "task-1245539" [ 1052.395591] env[63418]: _type = "Task" [ 1052.395591] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.402339] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245539, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.747957] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.778632] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79aeb2a-679b-4ea0-889e-938cb327455b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.804136] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 0 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.808397] env[63418]: DEBUG nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updated inventory for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with generation 142 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1052.808658] env[63418]: DEBUG nova.compute.provider_tree [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating resource provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c generation from 142 to 143 during operation: update_inventory {{(pid=63418) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1052.808836] env[63418]: DEBUG nova.compute.provider_tree [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Updating inventory in ProviderTree for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1052.905500] env[63418]: DEBUG oslo_vmware.api [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245539, 'name': PowerOnVM_Task, 'duration_secs': 0.403602} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.905823] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.906067] env[63418]: INFO nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Took 7.94 seconds to spawn the instance on the hypervisor. [ 1052.906277] env[63418]: DEBUG nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1052.907123] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21c8d25-1db4-43b3-bdb6-f51711c41dd3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.313774] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.314482] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.040s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.315036] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1053.317510] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-398698b1-4219-4196-b1e2-28a805904a7f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.319297] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.021s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.320631] env[63418]: INFO nova.compute.claims [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.328247] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1053.328247] env[63418]: value = "task-1245540" [ 1053.328247] env[63418]: _type = "Task" [ 1053.328247] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.335959] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245540, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.423850] env[63418]: INFO nova.compute.manager [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Took 12.74 seconds to build instance. [ 1053.825545] env[63418]: DEBUG nova.compute.utils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.829043] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1053.829211] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1053.840128] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245540, 'name': PowerOffVM_Task, 'duration_secs': 0.239008} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.841218] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.841415] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 17 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.909162] env[63418]: DEBUG nova.policy [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc5a12be9ce64b5294d8aab8af52e61c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c585aa2fd2142aba53f8d5a91f5e332', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1053.925617] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a294612a-a023-4d74-9aea-9d5d95eedc3f tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "fe38702e-953d-4748-a523-401e798b2271" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.247s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.330066] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1054.347389] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.347875] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.347875] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.347973] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.352023] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.354646] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05a0a15f-f526-485e-b09b-440f939549f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.371527] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1054.371527] env[63418]: value = "task-1245541" [ 1054.371527] env[63418]: _type = "Task" [ 1054.371527] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.381735] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.409147] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Successfully created port: 7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.524942] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a010ae-dbac-4c4e-ad7c-50cd92c6b5dc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.534210] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8f69a2-0f38-4fa2-885e-d720a8e3842b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.569092] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e7f3ab-30bc-4594-b078-4ba541eda2f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.577305] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5229c97-b54e-4f8b-a5cd-a2eaf6d06133 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.591171] env[63418]: DEBUG nova.compute.provider_tree [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.722254] env[63418]: DEBUG nova.compute.manager [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Received event network-changed-9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1054.722416] env[63418]: DEBUG nova.compute.manager [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Refreshing instance network info cache due to event network-changed-9e929283-d11e-4069-ab19-e7195f2cfd1b. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1054.722635] env[63418]: DEBUG oslo_concurrency.lockutils [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] Acquiring lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.722839] env[63418]: DEBUG oslo_concurrency.lockutils [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] Acquired lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.723016] env[63418]: DEBUG nova.network.neutron [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Refreshing network info cache for port 9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1054.886103] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245541, 'name': ReconfigVM_Task, 'duration_secs': 0.155707} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.886103] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 33 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1055.094663] env[63418]: DEBUG nova.scheduler.client.report [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1055.340695] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1055.371144] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.371471] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.371677] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.371888] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.372079] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.377333] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.377333] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.377333] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.377333] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.377333] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.377333] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.378090] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c90de3-6efb-43da-8227-8b79a52859e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.400070] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.400490] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.400835] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.401184] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.401426] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.401653] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.402577] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.402852] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.403158] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.403401] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.403675] env[63418]: DEBUG nova.virt.hardware [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.418904] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1055.420080] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a2d060f-fa69-44b7-94db-e97d4e7023b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.445651] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5822a3-55aa-4444-be59-9f35bef0cb3b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.449629] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1055.449629] env[63418]: value = "task-1245542" [ 1055.449629] env[63418]: _type = "Task" [ 1055.449629] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.466214] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245542, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.509247] env[63418]: DEBUG nova.network.neutron [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Updated VIF entry in instance network info cache for port 9e929283-d11e-4069-ab19-e7195f2cfd1b. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1055.509684] env[63418]: DEBUG nova.network.neutron [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Updating instance_info_cache with network_info: [{"id": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "address": "fa:16:3e:4b:1f:84", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e929283-d1", "ovs_interfaceid": "9e929283-d11e-4069-ab19-e7195f2cfd1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.601846] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.602071] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1055.604576] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.857s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.604801] env[63418]: DEBUG nova.objects.instance [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid 5673b437-e993-47f8-8eaf-4be99869bfc6 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.959695] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245542, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.994321] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Successfully updated port: 7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.012866] env[63418]: DEBUG oslo_concurrency.lockutils [req-32a9cc44-7d30-42b6-bebb-ab210774b250 req-21776bda-f5f1-47f3-b2d9-284f2ae7d7b8 service nova] Releasing lock "refresh_cache-fe38702e-953d-4748-a523-401e798b2271" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.107800] env[63418]: DEBUG nova.compute.utils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1056.111920] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1056.112101] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1056.190859] env[63418]: DEBUG nova.policy [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc5a12be9ce64b5294d8aab8af52e61c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c585aa2fd2142aba53f8d5a91f5e332', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.254884] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466d38d0-c283-4412-9f58-53226bf435fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.262606] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0f0c8a-7e66-4236-800c-72420d2df774 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.293078] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16d803f-d2c4-40bd-a654-ff24678f685a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.300653] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784f1a32-15b7-45b4-9bf7-12527a2f76c7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.313477] env[63418]: DEBUG nova.compute.provider_tree [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.437349] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Successfully created port: cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.460382] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245542, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.497263] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "refresh_cache-960184bb-0e6f-44ab-a119-fb43c27df929" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.497406] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "refresh_cache-960184bb-0e6f-44ab-a119-fb43c27df929" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.497560] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.612639] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1056.747476] env[63418]: DEBUG nova.compute.manager [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Received event network-vif-plugged-7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1056.747753] env[63418]: DEBUG oslo_concurrency.lockutils [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] Acquiring lock "960184bb-0e6f-44ab-a119-fb43c27df929-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.748047] env[63418]: DEBUG oslo_concurrency.lockutils [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] Lock "960184bb-0e6f-44ab-a119-fb43c27df929-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.748107] env[63418]: DEBUG oslo_concurrency.lockutils [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] Lock "960184bb-0e6f-44ab-a119-fb43c27df929-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.748263] env[63418]: DEBUG nova.compute.manager [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] No waiting events found dispatching network-vif-plugged-7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1056.748434] env[63418]: WARNING nova.compute.manager [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Received unexpected event network-vif-plugged-7e337af0-6e03-41d4-9dc2-cc843a9ba7aa for instance with vm_state building and task_state spawning. [ 1056.748594] env[63418]: DEBUG nova.compute.manager [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Received event network-changed-7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1056.748751] env[63418]: DEBUG nova.compute.manager [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Refreshing instance network info cache due to event network-changed-7e337af0-6e03-41d4-9dc2-cc843a9ba7aa. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1056.748991] env[63418]: DEBUG oslo_concurrency.lockutils [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] Acquiring lock "refresh_cache-960184bb-0e6f-44ab-a119-fb43c27df929" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.815997] env[63418]: DEBUG nova.scheduler.client.report [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1056.960910] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245542, 'name': ReconfigVM_Task, 'duration_secs': 1.193193} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.961273] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1056.962083] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ecbdd8-046a-4fea-a14d-73380e599860 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.986782] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.987089] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6240b738-b5c3-4b1f-9db7-b90bd5dab65e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.005727] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1057.005727] env[63418]: value = "task-1245543" [ 1057.005727] env[63418]: _type = "Task" [ 1057.005727] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.015869] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245543, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.036443] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1057.171518] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Updating instance_info_cache with network_info: [{"id": "7e337af0-6e03-41d4-9dc2-cc843a9ba7aa", "address": "fa:16:3e:25:5e:17", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e337af0-6e", "ovs_interfaceid": "7e337af0-6e03-41d4-9dc2-cc843a9ba7aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.210363] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.210502] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Cleaning up deleted instances with incomplete migration {{(pid=63418) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11616}} [ 1057.323011] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.345791] env[63418]: INFO nova.scheduler.client.report [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance 5673b437-e993-47f8-8eaf-4be99869bfc6 [ 1057.516751] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.622701] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1057.642167] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.642458] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.642608] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.642793] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.642958] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.643166] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.643373] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.643565] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.643751] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.643916] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.644098] env[63418]: DEBUG nova.virt.hardware [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.644968] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f890bb3-46fa-4083-b18f-6ac081efb901 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.653092] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9abbee1-17fb-4d47-95a8-84647f90492a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.675140] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "refresh_cache-960184bb-0e6f-44ab-a119-fb43c27df929" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.675465] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Instance network_info: |[{"id": "7e337af0-6e03-41d4-9dc2-cc843a9ba7aa", "address": "fa:16:3e:25:5e:17", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e337af0-6e", "ovs_interfaceid": "7e337af0-6e03-41d4-9dc2-cc843a9ba7aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1057.675762] env[63418]: DEBUG oslo_concurrency.lockutils [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] Acquired lock "refresh_cache-960184bb-0e6f-44ab-a119-fb43c27df929" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.675980] env[63418]: DEBUG nova.network.neutron [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Refreshing network info cache for port 7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.677819] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:5e:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e337af0-6e03-41d4-9dc2-cc843a9ba7aa', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.685751] env[63418]: DEBUG oslo.service.loopingcall [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.686696] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1057.686944] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f788da8e-5977-4804-8e17-fed5ec53b8cd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.706967] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.706967] env[63418]: value = "task-1245544" [ 1057.706967] env[63418]: _type = "Task" [ 1057.706967] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.715554] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245544, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.854625] env[63418]: DEBUG oslo_concurrency.lockutils [None req-397250f5-3302-47c0-9c2e-c36631395e17 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "5673b437-e993-47f8-8eaf-4be99869bfc6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.532s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.017117] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245543, 'name': ReconfigVM_Task, 'duration_secs': 0.960914} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.017527] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.017664] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 50 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.218855] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Successfully updated port: cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.225405] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245544, 'name': CreateVM_Task, 'duration_secs': 0.314944} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.225610] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.226428] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.226534] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.226908] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.227213] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63107253-a082-44e3-ab99-0adaa5785b2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.231872] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1058.231872] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5287b3a0-13ed-2ec8-5715-3641dcca52cc" [ 1058.231872] env[63418]: _type = "Task" [ 1058.231872] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.241158] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5287b3a0-13ed-2ec8-5715-3641dcca52cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.393059] env[63418]: DEBUG nova.network.neutron [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Updated VIF entry in instance network info cache for port 7e337af0-6e03-41d4-9dc2-cc843a9ba7aa. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.393466] env[63418]: DEBUG nova.network.neutron [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Updating instance_info_cache with network_info: [{"id": "7e337af0-6e03-41d4-9dc2-cc843a9ba7aa", "address": "fa:16:3e:25:5e:17", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e337af0-6e", "ovs_interfaceid": "7e337af0-6e03-41d4-9dc2-cc843a9ba7aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.524012] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8dda70-d5dd-478e-8a7b-636051a480d3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.547499] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7f3039-50ef-46ec-87b4-7ba5b3ea28bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.566834] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 67 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.720609] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "refresh_cache-1e6828a8-02ea-434f-b587-8d4c6854d944" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.720777] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "refresh_cache-1e6828a8-02ea-434f-b587-8d4c6854d944" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.720931] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.742933] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5287b3a0-13ed-2ec8-5715-3641dcca52cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010949} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.743234] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.743472] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.743721] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.743871] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.744066] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.744310] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19382512-8326-438d-972c-7c51953752f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.753891] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.754087] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.754799] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4c94d53-c7ac-4c61-a7f2-1fe17d3aba95 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.759755] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1058.759755] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a291fb-b47c-dbd2-c140-a25843820042" [ 1058.759755] env[63418]: _type = "Task" [ 1058.759755] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.767675] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a291fb-b47c-dbd2-c140-a25843820042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.775900] env[63418]: DEBUG nova.compute.manager [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Received event network-vif-plugged-cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1058.776141] env[63418]: DEBUG oslo_concurrency.lockutils [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] Acquiring lock "1e6828a8-02ea-434f-b587-8d4c6854d944-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.776976] env[63418]: DEBUG oslo_concurrency.lockutils [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.776976] env[63418]: DEBUG oslo_concurrency.lockutils [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.776976] env[63418]: DEBUG nova.compute.manager [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] No waiting events found dispatching network-vif-plugged-cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1058.776976] env[63418]: WARNING nova.compute.manager [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Received unexpected event network-vif-plugged-cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 for instance with vm_state building and task_state spawning. [ 1058.777398] env[63418]: DEBUG nova.compute.manager [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Received event network-changed-cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1058.777398] env[63418]: DEBUG nova.compute.manager [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Refreshing instance network info cache due to event network-changed-cb2ab1f6-383f-40c9-ba8a-059ff9e138d9. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1058.777398] env[63418]: DEBUG oslo_concurrency.lockutils [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] Acquiring lock "refresh_cache-1e6828a8-02ea-434f-b587-8d4c6854d944" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.846605] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.846863] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.896322] env[63418]: DEBUG oslo_concurrency.lockutils [req-937f00fa-e504-43c6-bbff-b3abb8edf657 req-ca326430-72cb-4495-bdce-fedea6acd99e service nova] Releasing lock "refresh_cache-960184bb-0e6f-44ab-a119-fb43c27df929" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.253196] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1059.270136] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52a291fb-b47c-dbd2-c140-a25843820042, 'name': SearchDatastore_Task, 'duration_secs': 0.01394} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.271023] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd349a3d-fce8-4aeb-8daa-b130ee121a3e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.278164] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1059.278164] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52550460-ed39-97b4-8de3-44851099083a" [ 1059.278164] env[63418]: _type = "Task" [ 1059.278164] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.285684] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52550460-ed39-97b4-8de3-44851099083a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.349408] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1059.412244] env[63418]: DEBUG nova.network.neutron [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Updating instance_info_cache with network_info: [{"id": "cb2ab1f6-383f-40c9-ba8a-059ff9e138d9", "address": "fa:16:3e:5c:ac:35", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb2ab1f6-38", "ovs_interfaceid": "cb2ab1f6-383f-40c9-ba8a-059ff9e138d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.715360] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.788527] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52550460-ed39-97b4-8de3-44851099083a, 'name': SearchDatastore_Task, 'duration_secs': 0.020765} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.788845] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.789085] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 960184bb-0e6f-44ab-a119-fb43c27df929/960184bb-0e6f-44ab-a119-fb43c27df929.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.789348] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e871e7ec-044f-4d8f-bc78-dc338aa318e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.795367] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1059.795367] env[63418]: value = "task-1245545" [ 1059.795367] env[63418]: _type = "Task" [ 1059.795367] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.802659] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.873642] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.873981] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.875597] env[63418]: INFO nova.compute.claims [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.914193] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "refresh_cache-1e6828a8-02ea-434f-b587-8d4c6854d944" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.914496] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Instance network_info: |[{"id": "cb2ab1f6-383f-40c9-ba8a-059ff9e138d9", "address": "fa:16:3e:5c:ac:35", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb2ab1f6-38", "ovs_interfaceid": "cb2ab1f6-383f-40c9-ba8a-059ff9e138d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1059.914801] env[63418]: DEBUG oslo_concurrency.lockutils [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] Acquired lock "refresh_cache-1e6828a8-02ea-434f-b587-8d4c6854d944" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.914982] env[63418]: DEBUG nova.network.neutron [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Refreshing network info cache for port cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.916115] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:ac:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb2ab1f6-383f-40c9-ba8a-059ff9e138d9', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.923507] env[63418]: DEBUG oslo.service.loopingcall [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.926571] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.927042] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff55c85c-71a9-4130-9e6e-71bc3bad9464 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.949486] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.949486] env[63418]: value = "task-1245546" [ 1059.949486] env[63418]: _type = "Task" [ 1059.949486] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.957824] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245546, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.185453] env[63418]: DEBUG nova.network.neutron [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Updated VIF entry in instance network info cache for port cb2ab1f6-383f-40c9-ba8a-059ff9e138d9. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.186330] env[63418]: DEBUG nova.network.neutron [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Updating instance_info_cache with network_info: [{"id": "cb2ab1f6-383f-40c9-ba8a-059ff9e138d9", "address": "fa:16:3e:5c:ac:35", "network": {"id": "2ad9f3fe-3fa9-4f10-8c81-6600632dba68", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1458323748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c585aa2fd2142aba53f8d5a91f5e332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb2ab1f6-38", "ovs_interfaceid": "cb2ab1f6-383f-40c9-ba8a-059ff9e138d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.205051] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.210500] env[63418]: DEBUG nova.network.neutron [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Port 6eb8c1f9-eec0-4763-8033-34a4efda65c1 binding to destination host cpu-1 is already ACTIVE {{(pid=63418) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1060.212408] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.212711] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Cleaning up deleted instances {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11578}} [ 1060.305259] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441002} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.305606] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 960184bb-0e6f-44ab-a119-fb43c27df929/960184bb-0e6f-44ab-a119-fb43c27df929.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.305833] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.306098] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dfd3845d-32ea-4119-869c-22da40838a65 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.315554] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1060.315554] env[63418]: value = "task-1245547" [ 1060.315554] env[63418]: _type = "Task" [ 1060.315554] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.325139] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.459772] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245546, 'name': CreateVM_Task, 'duration_secs': 0.422686} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.459930] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.460606] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.460781] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.461127] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.461393] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11707f35-79db-4ba6-82d8-52dfdcf0995d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.465621] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1060.465621] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]523d97e4-5d07-ce18-e4d3-393178385f10" [ 1060.465621] env[63418]: _type = "Task" [ 1060.465621] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.474825] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523d97e4-5d07-ce18-e4d3-393178385f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.689153] env[63418]: DEBUG oslo_concurrency.lockutils [req-97635caf-ccd0-49f2-8627-439369a3be60 req-dc6da72e-187c-4a17-8b5f-604bfcc67dd9 service nova] Releasing lock "refresh_cache-1e6828a8-02ea-434f-b587-8d4c6854d944" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.730523] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] There are 51 instances to clean {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11587}} [ 1060.730792] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 5673b437-e993-47f8-8eaf-4be99869bfc6] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1060.825658] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066478} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.825924] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1060.826730] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae753ac-3268-4417-9379-d43c7bb31c2b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.847784] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 960184bb-0e6f-44ab-a119-fb43c27df929/960184bb-0e6f-44ab-a119-fb43c27df929.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.848030] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a5da398-5a90-4a59-8273-bba5b4e0b708 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.865914] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1060.865914] env[63418]: value = "task-1245548" [ 1060.865914] env[63418]: _type = "Task" [ 1060.865914] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.872909] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245548, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.976130] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]523d97e4-5d07-ce18-e4d3-393178385f10, 'name': SearchDatastore_Task, 'duration_secs': 0.016667} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.978396] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.978639] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.978876] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.979036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.979226] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.979654] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c110a5e-9f7b-4a69-96b5-a503199e28e6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.995076] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.995273] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.996067] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1b0a96e-85a8-44a8-ac94-2a3705bf798a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.003529] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1061.003529] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e699e5-320f-d402-b842-fec8d67d925a" [ 1061.003529] env[63418]: _type = "Task" [ 1061.003529] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.010722] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e699e5-320f-d402-b842-fec8d67d925a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.021872] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa8c92a-84f8-4bdf-b7d8-44ec482c857c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.028128] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d03e354-39a6-4115-a7bf-68945477628b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.058599] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f894a90-0cd2-473f-8e74-971f0c79a561 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.064972] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd2591b-5418-48fc-85ae-fb733ee15b30 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.713015] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 2dae6f5b-9bdc-4862-8654-2684ea22c6ac] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1061.729890] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.729890] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.729890] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.737466] env[63418]: DEBUG nova.compute.provider_tree [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.744411] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245548, 'name': ReconfigVM_Task, 'duration_secs': 0.549811} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.748367] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 960184bb-0e6f-44ab-a119-fb43c27df929/960184bb-0e6f-44ab-a119-fb43c27df929.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.749280] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52e699e5-320f-d402-b842-fec8d67d925a, 'name': SearchDatastore_Task, 'duration_secs': 0.066023} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.749497] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79740f0a-a6a6-49f2-bf46-325b4b17ea10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.751887] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-113bba60-15fe-4b1e-b147-9e8e61daeb21 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.758039] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1061.758039] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52555882-0def-ddd3-e750-db544bbd34b7" [ 1061.758039] env[63418]: _type = "Task" [ 1061.758039] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.759272] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1061.759272] env[63418]: value = "task-1245549" [ 1061.759272] env[63418]: _type = "Task" [ 1061.759272] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.769378] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52555882-0def-ddd3-e750-db544bbd34b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.772739] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245549, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.218518] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d0633a7d-33b4-4613-9501-81cef2ab3680] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1062.241332] env[63418]: DEBUG nova.scheduler.client.report [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1062.270482] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52555882-0def-ddd3-e750-db544bbd34b7, 'name': SearchDatastore_Task, 'duration_secs': 0.013668} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.273339] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.273629] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1e6828a8-02ea-434f-b587-8d4c6854d944/1e6828a8-02ea-434f-b587-8d4c6854d944.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.273900] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245549, 'name': Rename_Task, 'duration_secs': 0.175807} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.274325] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8450fbe7-bf37-4bdf-8345-db79e721dc56 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.276035] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1062.276256] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dadd8d9c-db1c-43f1-bf9c-d1b6e623dd8c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.281430] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1062.281430] env[63418]: value = "task-1245550" [ 1062.281430] env[63418]: _type = "Task" [ 1062.281430] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.282568] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1062.282568] env[63418]: value = "task-1245551" [ 1062.282568] env[63418]: _type = "Task" [ 1062.282568] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.292155] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245551, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.294857] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245550, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.722748] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 9ae2cb1e-3999-4471-8a81-7f86db857ff3] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1062.746428] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.872s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.747018] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1062.770229] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.770421] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.770590] env[63418]: DEBUG nova.network.neutron [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.794798] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245550, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448226} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.797979] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 1e6828a8-02ea-434f-b587-8d4c6854d944/1e6828a8-02ea-434f-b587-8d4c6854d944.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1062.798213] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.798730] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245551, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.798963] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d0459b7-1d3a-4194-a79f-a9356002ee54 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.804471] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1062.804471] env[63418]: value = "task-1245552" [ 1062.804471] env[63418]: _type = "Task" [ 1062.804471] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.811436] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245552, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.226525] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 0963bddc-6d3d-424d-a218-a2341fa9dab7] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1063.251894] env[63418]: DEBUG nova.compute.utils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.253325] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1063.253830] env[63418]: DEBUG nova.network.neutron [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.291178] env[63418]: DEBUG nova.policy [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '09fd80e0c51e4f17990f42421cc1b49e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4dac4885b78c414bbe379918a0d8cf61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1063.297640] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245551, 'name': PowerOnVM_Task, 'duration_secs': 0.640794} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.297898] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.298110] env[63418]: INFO nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Took 7.96 seconds to spawn the instance on the hypervisor. [ 1063.298297] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1063.299063] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a575f02-b266-4742-8917-7631e887c738 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.315158] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245552, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059298} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.315384] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.316136] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2dcd26f-caba-4006-a72c-576ede9cc94b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.338826] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 1e6828a8-02ea-434f-b587-8d4c6854d944/1e6828a8-02ea-434f-b587-8d4c6854d944.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.342236] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e422c10-2535-4d84-91f0-8b7659fc2aa8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.362753] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1063.362753] env[63418]: value = "task-1245553" [ 1063.362753] env[63418]: _type = "Task" [ 1063.362753] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.370453] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245553, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.508663] env[63418]: DEBUG nova.network.neutron [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.530084] env[63418]: DEBUG nova.network.neutron [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Successfully created port: d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.730236] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 3e645774-d1b3-468b-86a5-ff0c07e77f4e] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1063.757175] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1063.818642] env[63418]: INFO nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Took 13.57 seconds to build instance. [ 1063.875606] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245553, 'name': ReconfigVM_Task, 'duration_secs': 0.307473} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.875946] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 1e6828a8-02ea-434f-b587-8d4c6854d944/1e6828a8-02ea-434f-b587-8d4c6854d944.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.876654] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20097ab6-489e-4e2a-bb1c-a8c2f5843df7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.882790] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1063.882790] env[63418]: value = "task-1245554" [ 1063.882790] env[63418]: _type = "Task" [ 1063.882790] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.890774] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245554, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.011558] env[63418]: DEBUG oslo_concurrency.lockutils [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.233772] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 0c016d4e-ca34-4831-a567-e794012681db] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1064.320458] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "960184bb-0e6f-44ab-a119-fb43c27df929" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.078s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.392771] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245554, 'name': Rename_Task, 'duration_secs': 0.168805} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.393094] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.393336] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9cc5bb6-f00c-4002-834d-67c448eac29e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.398666] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1064.398666] env[63418]: value = "task-1245555" [ 1064.398666] env[63418]: _type = "Task" [ 1064.398666] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.406562] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.521058] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123cc83c-fa4a-4236-8f8a-059b891657ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.528265] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa8c190-25e8-4f18-935a-c6a5b934ec33 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.544095] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "80777e38-728b-4d2d-9003-f55ea7999299" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.544348] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "80777e38-728b-4d2d-9003-f55ea7999299" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.544583] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "80777e38-728b-4d2d-9003-f55ea7999299-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.544774] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "80777e38-728b-4d2d-9003-f55ea7999299-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.545155] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "80777e38-728b-4d2d-9003-f55ea7999299-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.547201] env[63418]: INFO nova.compute.manager [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Terminating instance [ 1064.736979] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 680c94e8-c6d9-4aab-92d4-047a0e70eec6] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1064.765565] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1064.793114] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.793459] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.793652] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.793861] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.794024] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.794186] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.794402] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.794572] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.794786] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.794956] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.795148] env[63418]: DEBUG nova.virt.hardware [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.796055] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43920cb1-960c-401b-a9f6-f3c433dda950 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.804251] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9914fef2-e145-45b5-b76f-2071b2983dee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.909386] env[63418]: DEBUG oslo_vmware.api [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245555, 'name': PowerOnVM_Task, 'duration_secs': 0.425385} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.909655] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.909852] env[63418]: INFO nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Took 7.29 seconds to spawn the instance on the hypervisor. [ 1064.910038] env[63418]: DEBUG nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1064.910790] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f106573e-6a8f-48d5-8379-653b0e03ee50 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.998473] env[63418]: DEBUG nova.compute.manager [req-0be04d18-4bc9-4c41-b4fe-9224a095abe9 req-91cb1ec9-405a-4503-9611-5526bb3997f1 service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Received event network-vif-plugged-d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1064.998729] env[63418]: DEBUG oslo_concurrency.lockutils [req-0be04d18-4bc9-4c41-b4fe-9224a095abe9 req-91cb1ec9-405a-4503-9611-5526bb3997f1 service nova] Acquiring lock "69e2b485-bb5d-4a90-8f78-66e6342b7928-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.998921] env[63418]: DEBUG oslo_concurrency.lockutils [req-0be04d18-4bc9-4c41-b4fe-9224a095abe9 req-91cb1ec9-405a-4503-9611-5526bb3997f1 service nova] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.999111] env[63418]: DEBUG oslo_concurrency.lockutils [req-0be04d18-4bc9-4c41-b4fe-9224a095abe9 req-91cb1ec9-405a-4503-9611-5526bb3997f1 service nova] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.999338] env[63418]: DEBUG nova.compute.manager [req-0be04d18-4bc9-4c41-b4fe-9224a095abe9 req-91cb1ec9-405a-4503-9611-5526bb3997f1 service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] No waiting events found dispatching network-vif-plugged-d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1064.999527] env[63418]: WARNING nova.compute.manager [req-0be04d18-4bc9-4c41-b4fe-9224a095abe9 req-91cb1ec9-405a-4503-9611-5526bb3997f1 service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Received unexpected event network-vif-plugged-d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 for instance with vm_state building and task_state spawning. [ 1065.051311] env[63418]: DEBUG nova.compute.manager [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1065.051543] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1065.052455] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95af4dbd-4fea-43e4-9b8b-5fb798f76f11 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.061204] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.061454] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abf6958e-721a-4dd0-8c20-5ddcd47ebcfa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.067140] env[63418]: DEBUG oslo_vmware.api [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1065.067140] env[63418]: value = "task-1245556" [ 1065.067140] env[63418]: _type = "Task" [ 1065.067140] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.074619] env[63418]: DEBUG oslo_vmware.api [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.092132] env[63418]: DEBUG nova.network.neutron [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Successfully updated port: d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.240337] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: b0887bcd-7ba7-4c0e-8d50-886e27d37649] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1065.426758] env[63418]: INFO nova.compute.manager [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Took 15.15 seconds to build instance. [ 1065.576950] env[63418]: DEBUG oslo_vmware.api [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245556, 'name': PowerOffVM_Task, 'duration_secs': 0.228982} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.577652] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.577755] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1065.578052] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a648791e-0abc-4e11-8dd4-451fa802ffba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.597483] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "refresh_cache-69e2b485-bb5d-4a90-8f78-66e6342b7928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.597587] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "refresh_cache-69e2b485-bb5d-4a90-8f78-66e6342b7928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.597764] env[63418]: DEBUG nova.network.neutron [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.655379] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1065.655379] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1065.655379] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Deleting the datastore file [datastore1] 80777e38-728b-4d2d-9003-f55ea7999299 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.655550] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-acb1e259-2b4e-4bf3-b8c3-66b7d831d520 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.662979] env[63418]: DEBUG oslo_vmware.api [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for the task: (returnval){ [ 1065.662979] env[63418]: value = "task-1245558" [ 1065.662979] env[63418]: _type = "Task" [ 1065.662979] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.672563] env[63418]: DEBUG oslo_vmware.api [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245558, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.692861] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24195f4-14c8-4fc9-88b9-4b543465cb75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.726615] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb026a8e-ded2-4916-ad27-0942fa3e6697 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.736820] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 83 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1065.744432] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 5dac16e3-06a9-443a-90c8-9aacdd23fd91] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1065.914640] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "960184bb-0e6f-44ab-a119-fb43c27df929" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.915039] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "960184bb-0e6f-44ab-a119-fb43c27df929" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.915298] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "960184bb-0e6f-44ab-a119-fb43c27df929-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.915493] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "960184bb-0e6f-44ab-a119-fb43c27df929-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.915703] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "960184bb-0e6f-44ab-a119-fb43c27df929-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.917842] env[63418]: INFO nova.compute.manager [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Terminating instance [ 1065.930063] env[63418]: DEBUG oslo_concurrency.lockutils [None req-5793d384-b46f-4792-ab5a-0c67e88eb201 tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.662s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.980429] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "1e6828a8-02ea-434f-b587-8d4c6854d944" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.980726] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.980959] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "1e6828a8-02ea-434f-b587-8d4c6854d944-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.981168] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.981343] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.983603] env[63418]: INFO nova.compute.manager [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Terminating instance [ 1066.130883] env[63418]: DEBUG nova.network.neutron [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.172757] env[63418]: DEBUG oslo_vmware.api [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Task: {'id': task-1245558, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16571} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.174997] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.175206] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1066.175386] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1066.175561] env[63418]: INFO nova.compute.manager [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1066.175837] env[63418]: DEBUG oslo.service.loopingcall [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.176048] env[63418]: DEBUG nova.compute.manager [-] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1066.176143] env[63418]: DEBUG nova.network.neutron [-] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1066.244431] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1066.244735] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed20ad39-8314-45fd-a6af-751a639ebce6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.246996] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 23af1d2e-f8ec-4c09-a070-9719b593f83a] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1066.253201] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1066.253201] env[63418]: value = "task-1245559" [ 1066.253201] env[63418]: _type = "Task" [ 1066.253201] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.261984] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.277904] env[63418]: DEBUG nova.network.neutron [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Updating instance_info_cache with network_info: [{"id": "d3da5ac7-6b95-4055-afd8-9e8f5a2d9657", "address": "fa:16:3e:93:6e:aa", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3da5ac7-6b", "ovs_interfaceid": "d3da5ac7-6b95-4055-afd8-9e8f5a2d9657", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.421771] env[63418]: DEBUG nova.compute.manager [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1066.421997] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.422899] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d1e248-3587-4cc8-bcec-ed886b0175bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.430383] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.430708] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1dbded9-e538-4693-abea-6fb752b13590 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.438064] env[63418]: DEBUG oslo_vmware.api [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1066.438064] env[63418]: value = "task-1245560" [ 1066.438064] env[63418]: _type = "Task" [ 1066.438064] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.446210] env[63418]: DEBUG oslo_vmware.api [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.488387] env[63418]: DEBUG nova.compute.manager [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1066.488387] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.488885] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec236dfc-f2dc-4ee0-81b4-0162475589ab {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.497623] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.497963] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d91fc8a-7c5b-424e-8640-fc933487bad5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.504478] env[63418]: DEBUG oslo_vmware.api [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1066.504478] env[63418]: value = "task-1245561" [ 1066.504478] env[63418]: _type = "Task" [ 1066.504478] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.513166] env[63418]: DEBUG oslo_vmware.api [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245561, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.687418] env[63418]: DEBUG nova.compute.manager [req-a4170c19-e0b9-4af9-a4f5-85109da30b0a req-28f04455-94b6-42f0-8b72-01a86d30a022 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Received event network-vif-deleted-0079f927-f058-4093-b9d5-7950de3da99a {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1066.687664] env[63418]: INFO nova.compute.manager [req-a4170c19-e0b9-4af9-a4f5-85109da30b0a req-28f04455-94b6-42f0-8b72-01a86d30a022 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Neutron deleted interface 0079f927-f058-4093-b9d5-7950de3da99a; detaching it from the instance and deleting it from the info cache [ 1066.687842] env[63418]: DEBUG nova.network.neutron [req-a4170c19-e0b9-4af9-a4f5-85109da30b0a req-28f04455-94b6-42f0-8b72-01a86d30a022 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.750276] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 26147513-3b25-4cc1-991a-34a724f73711] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1066.764795] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245559, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.780579] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "refresh_cache-69e2b485-bb5d-4a90-8f78-66e6342b7928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.780951] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Instance network_info: |[{"id": "d3da5ac7-6b95-4055-afd8-9e8f5a2d9657", "address": "fa:16:3e:93:6e:aa", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3da5ac7-6b", "ovs_interfaceid": "d3da5ac7-6b95-4055-afd8-9e8f5a2d9657", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1066.781747] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:6e:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3da5ac7-6b95-4055-afd8-9e8f5a2d9657', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.791717] env[63418]: DEBUG oslo.service.loopingcall [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.792236] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.792489] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c62f9be-1c09-4651-aa4f-d92c1bb4b088 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.813693] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.813693] env[63418]: value = "task-1245562" [ 1066.813693] env[63418]: _type = "Task" [ 1066.813693] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.822612] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245562, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.948472] env[63418]: DEBUG oslo_vmware.api [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245560, 'name': PowerOffVM_Task, 'duration_secs': 0.237571} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.948886] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1066.949228] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1066.949818] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f24cf28-64c7-40fc-b1f2-00062a6f9fc0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.019035] env[63418]: DEBUG oslo_vmware.api [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245561, 'name': PowerOffVM_Task, 'duration_secs': 0.209702} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.021142] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.021142] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.021915] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7724de3b-f899-465c-b321-a5e7ec8db01a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.026220] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.026220] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.026220] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleting the datastore file [datastore2] 960184bb-0e6f-44ab-a119-fb43c27df929 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.026220] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58c37f4e-26f5-4a14-b546-f1e61f8e6b00 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.033911] env[63418]: DEBUG oslo_vmware.api [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1067.033911] env[63418]: value = "task-1245565" [ 1067.033911] env[63418]: _type = "Task" [ 1067.033911] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.036876] env[63418]: DEBUG nova.compute.manager [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Received event network-changed-d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1067.037162] env[63418]: DEBUG nova.compute.manager [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Refreshing instance network info cache due to event network-changed-d3da5ac7-6b95-4055-afd8-9e8f5a2d9657. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1067.037469] env[63418]: DEBUG oslo_concurrency.lockutils [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] Acquiring lock "refresh_cache-69e2b485-bb5d-4a90-8f78-66e6342b7928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.037716] env[63418]: DEBUG oslo_concurrency.lockutils [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] Acquired lock "refresh_cache-69e2b485-bb5d-4a90-8f78-66e6342b7928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.037978] env[63418]: DEBUG nova.network.neutron [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Refreshing network info cache for port d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.050465] env[63418]: DEBUG oslo_vmware.api [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245565, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.111552] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.111841] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.112080] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleting the datastore file [datastore2] 1e6828a8-02ea-434f-b587-8d4c6854d944 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.112391] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67456829-6189-4ba1-84b3-d913886289cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.120015] env[63418]: DEBUG oslo_vmware.api [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for the task: (returnval){ [ 1067.120015] env[63418]: value = "task-1245566" [ 1067.120015] env[63418]: _type = "Task" [ 1067.120015] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.128941] env[63418]: DEBUG oslo_vmware.api [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.164453] env[63418]: DEBUG nova.network.neutron [-] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.194170] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0189d133-142f-49f4-aa43-8fba947809e0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.203275] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7176a94-4e3c-4e32-a94c-cdf62865ded0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.230448] env[63418]: DEBUG nova.compute.manager [req-a4170c19-e0b9-4af9-a4f5-85109da30b0a req-28f04455-94b6-42f0-8b72-01a86d30a022 service nova] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Detach interface failed, port_id=0079f927-f058-4093-b9d5-7950de3da99a, reason: Instance 80777e38-728b-4d2d-9003-f55ea7999299 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1067.253900] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 0f99b32a-0125-4df0-919c-e5456b4ae4a5] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1067.264202] env[63418]: DEBUG oslo_vmware.api [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245559, 'name': PowerOnVM_Task, 'duration_secs': 0.57514} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.264476] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.264665] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-deaf6322-54bf-410e-971f-8064b3ce09a4 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance '5a8b347c-e9ec-4355-bcf7-b7d205da194c' progress to 100 {{(pid=63418) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.334346] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245562, 'name': CreateVM_Task, 'duration_secs': 0.372263} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.334491] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.335204] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.335379] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.335739] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.336245] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7dd1116-89a2-40f1-9573-5c6b7b9619af {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.341044] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1067.341044] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]528d7d56-c030-4349-4c0f-1ea50fcb77b9" [ 1067.341044] env[63418]: _type = "Task" [ 1067.341044] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.349216] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528d7d56-c030-4349-4c0f-1ea50fcb77b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.546510] env[63418]: DEBUG oslo_vmware.api [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245565, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141253} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.546855] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.547059] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.547240] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.547417] env[63418]: INFO nova.compute.manager [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1067.547657] env[63418]: DEBUG oslo.service.loopingcall [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.548201] env[63418]: DEBUG nova.compute.manager [-] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1067.548305] env[63418]: DEBUG nova.network.neutron [-] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.633496] env[63418]: DEBUG oslo_vmware.api [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Task: {'id': task-1245566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135316} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.633773] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.633963] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.634156] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.634508] env[63418]: INFO nova.compute.manager [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1067.634778] env[63418]: DEBUG oslo.service.loopingcall [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.634985] env[63418]: DEBUG nova.compute.manager [-] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1067.635087] env[63418]: DEBUG nova.network.neutron [-] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.666961] env[63418]: INFO nova.compute.manager [-] [instance: 80777e38-728b-4d2d-9003-f55ea7999299] Took 1.49 seconds to deallocate network for instance. [ 1067.760297] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 4838d2b2-a187-48ae-a3b1-8ca24b961359] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1067.852191] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]528d7d56-c030-4349-4c0f-1ea50fcb77b9, 'name': SearchDatastore_Task, 'duration_secs': 0.00908} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.852505] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.852750] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.852987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.853150] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.853505] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.853650] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d36e44d7-31c2-4143-ba85-f7311ac490b7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.865515] env[63418]: DEBUG nova.network.neutron [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Updated VIF entry in instance network info cache for port d3da5ac7-6b95-4055-afd8-9e8f5a2d9657. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.865515] env[63418]: DEBUG nova.network.neutron [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Updating instance_info_cache with network_info: [{"id": "d3da5ac7-6b95-4055-afd8-9e8f5a2d9657", "address": "fa:16:3e:93:6e:aa", "network": {"id": "eb6a2877-cf91-4e2f-9f32-458d8211adcc", "bridge": "br-int", "label": "tempest-ServersTestJSON-197643439-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4dac4885b78c414bbe379918a0d8cf61", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3da5ac7-6b", "ovs_interfaceid": "d3da5ac7-6b95-4055-afd8-9e8f5a2d9657", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.868450] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.868658] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.870066] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd8e0f88-f1f5-411e-9815-b9fde237e775 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.875398] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1067.875398] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5257e1bb-8297-d5fb-0578-6c3674499729" [ 1067.875398] env[63418]: _type = "Task" [ 1067.875398] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.886673] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5257e1bb-8297-d5fb-0578-6c3674499729, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.173345] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.173751] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.174071] env[63418]: DEBUG nova.objects.instance [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lazy-loading 'resources' on Instance uuid 80777e38-728b-4d2d-9003-f55ea7999299 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.263236] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 24a97d0a-252b-4bbd-9dfc-57767ef8014d] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1068.371085] env[63418]: DEBUG oslo_concurrency.lockutils [req-7e0e6345-3f19-4af2-a29f-3f0913e5bdf0 req-df65151d-67ad-4872-8fe3-0b1ad5f74c1a service nova] Releasing lock "refresh_cache-69e2b485-bb5d-4a90-8f78-66e6342b7928" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.385722] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5257e1bb-8297-d5fb-0578-6c3674499729, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.386473] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d155fd99-fe06-46fc-b8b7-fed5320f8f89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.392509] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1068.392509] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bd30fa-fbff-faf0-f924-e0cf9d3af060" [ 1068.392509] env[63418]: _type = "Task" [ 1068.392509] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.396756] env[63418]: DEBUG nova.network.neutron [-] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.400582] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bd30fa-fbff-faf0-f924-e0cf9d3af060, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.465438] env[63418]: DEBUG nova.network.neutron [-] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.711406] env[63418]: DEBUG nova.compute.manager [req-72557514-4f55-4b21-a92d-2824f28e6208 req-9c510b5b-a9ef-4e9d-a53c-219f9e2d56d2 service nova] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Received event network-vif-deleted-cb2ab1f6-383f-40c9-ba8a-059ff9e138d9 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1068.766839] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 14832e8e-3b4a-462e-84bb-d5a1e541d329] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1068.905012] env[63418]: INFO nova.compute.manager [-] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Took 1.36 seconds to deallocate network for instance. [ 1068.905427] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52bd30fa-fbff-faf0-f924-e0cf9d3af060, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.907326] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.907594] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 69e2b485-bb5d-4a90-8f78-66e6342b7928/69e2b485-bb5d-4a90-8f78-66e6342b7928.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.910880] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-556424cf-2ae6-47e8-8a4d-21e8e8f31f01 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.921758] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1068.921758] env[63418]: value = "task-1245567" [ 1068.921758] env[63418]: _type = "Task" [ 1068.921758] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.927347] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0d41ff-e1c6-43f7-a96a-3b6f86a87d00 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.932876] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.938333] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0ef499-2b76-454c-8dc6-e0e5c9c61f75 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.973284] env[63418]: INFO nova.compute.manager [-] [instance: 1e6828a8-02ea-434f-b587-8d4c6854d944] Took 1.34 seconds to deallocate network for instance. [ 1068.976312] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d48990-7de5-42c0-8a05-4361b835f3f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.987386] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778e8720-d017-4ede-83e5-6d29d7ca4f76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.003174] env[63418]: DEBUG nova.compute.provider_tree [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.068269] env[63418]: DEBUG nova.compute.manager [req-ca76b7d2-13e8-4e43-baee-02d5bc999214 req-50b719af-0854-4a65-bb19-6b0895fad87c service nova] [instance: 960184bb-0e6f-44ab-a119-fb43c27df929] Received event network-vif-deleted-7e337af0-6e03-41d4-9dc2-cc843a9ba7aa {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1069.270932] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 039f4cf6-137b-4356-a1b8-dd47f8f0d6f5] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1069.417121] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.431584] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465796} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.431726] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore1] 69e2b485-bb5d-4a90-8f78-66e6342b7928/69e2b485-bb5d-4a90-8f78-66e6342b7928.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.431937] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.432211] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d56806f-5242-4206-92fc-3e3d9c86b4f9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.438931] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1069.438931] env[63418]: value = "task-1245568" [ 1069.438931] env[63418]: _type = "Task" [ 1069.438931] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.446450] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.483332] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.506750] env[63418]: DEBUG nova.scheduler.client.report [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1069.774522] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 054bdcde-55a6-4f86-a230-92b2115e47b0] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1069.814007] env[63418]: DEBUG nova.network.neutron [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Port 6eb8c1f9-eec0-4763-8033-34a4efda65c1 binding to destination host cpu-1 is already ACTIVE {{(pid=63418) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1069.814296] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.814453] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.814617] env[63418]: DEBUG nova.network.neutron [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.948610] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07383} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.948878] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1069.949730] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7361a3bb-7a81-4278-930f-618793905318 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.973218] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 69e2b485-bb5d-4a90-8f78-66e6342b7928/69e2b485-bb5d-4a90-8f78-66e6342b7928.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.973497] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6285e61-b611-4aa6-83d2-f0c07916bf7a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.994286] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1069.994286] env[63418]: value = "task-1245569" [ 1069.994286] env[63418]: _type = "Task" [ 1069.994286] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.003646] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245569, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.011483] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.013641] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.597s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.014029] env[63418]: DEBUG nova.objects.instance [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lazy-loading 'resources' on Instance uuid 960184bb-0e6f-44ab-a119-fb43c27df929 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.033179] env[63418]: INFO nova.scheduler.client.report [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Deleted allocations for instance 80777e38-728b-4d2d-9003-f55ea7999299 [ 1070.277723] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: dcb52a03-8f82-4b33-8dc6-3924140db8d3] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1070.505843] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245569, 'name': ReconfigVM_Task, 'duration_secs': 0.278902} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.506510] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 69e2b485-bb5d-4a90-8f78-66e6342b7928/69e2b485-bb5d-4a90-8f78-66e6342b7928.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.507312] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbde22ad-ec21-497e-bdce-fd046b255773 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.514439] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1070.514439] env[63418]: value = "task-1245570" [ 1070.514439] env[63418]: _type = "Task" [ 1070.514439] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.526410] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245570, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.527390] env[63418]: DEBUG nova.network.neutron [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.542191] env[63418]: DEBUG oslo_concurrency.lockutils [None req-52919a5e-454f-4fd9-a323-aac9de3cb9f1 tempest-ServersV294TestFqdnHostnames-1362782648 tempest-ServersV294TestFqdnHostnames-1362782648-project-member] Lock "80777e38-728b-4d2d-9003-f55ea7999299" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.998s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.650806] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67b74ed-df93-47b6-b4f9-19d75229f2ac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.658592] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fe8c1f-5345-4eb1-bb05-41ef18f58bb7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.687043] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09fbcb8-7eec-4845-a9de-2d5712a13f02 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.694341] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebc4d90-cb36-4428-8b59-7d5bd2d6da02 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.708302] env[63418]: DEBUG nova.compute.provider_tree [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.780529] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: bb9fe6f5-6908-4a0d-862f-9d2a9acd54f5] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1071.023741] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245570, 'name': Rename_Task, 'duration_secs': 0.138629} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.024108] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.024358] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b97b06c1-6fc6-4ca7-9639-e65aa63388cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.029541] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.031841] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1071.031841] env[63418]: value = "task-1245571" [ 1071.031841] env[63418]: _type = "Task" [ 1071.031841] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.039478] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245571, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.214194] env[63418]: DEBUG nova.scheduler.client.report [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1071.283666] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 1b1c8fca-dfcf-4b25-999d-5a83e58fd6f5] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1071.533658] env[63418]: DEBUG nova.compute.manager [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63418) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1071.547150] env[63418]: DEBUG oslo_vmware.api [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245571, 'name': PowerOnVM_Task, 'duration_secs': 0.483029} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.547685] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.547685] env[63418]: INFO nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Took 6.78 seconds to spawn the instance on the hypervisor. [ 1071.548027] env[63418]: DEBUG nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1071.548929] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a411d3c6-aa60-49d0-8e28-832b40602a24 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.719769] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.722302] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.239s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.722703] env[63418]: DEBUG nova.objects.instance [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lazy-loading 'resources' on Instance uuid 1e6828a8-02ea-434f-b587-8d4c6854d944 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.752450] env[63418]: INFO nova.scheduler.client.report [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted allocations for instance 960184bb-0e6f-44ab-a119-fb43c27df929 [ 1071.788770] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 586f51a0-90de-4f44-ac3e-758ceda9e316] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1072.066243] env[63418]: INFO nova.compute.manager [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Took 12.21 seconds to build instance. [ 1072.268318] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2a7b1dec-d87d-42d7-b0bf-a1c945d399ec tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "960184bb-0e6f-44ab-a119-fb43c27df929" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.353s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.290870] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 5ab04ddd-4cd6-4330-a9d7-081290c27586] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1072.346409] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bd6daf-da7c-4c76-9829-0138b51ca656 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.354391] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adddefc-2be3-416f-8c1c-9c989ff2bae6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.384346] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c969bc-06b3-4e8a-adb8-863cbaba98d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.392811] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8514b473-05ef-41e3-833a-0d5e87d5a9bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.407874] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.407874] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.408067] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "cbcbaf4f-ee24-4072-83ae-ffde59478928-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.409022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.409022] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.410174] env[63418]: DEBUG nova.compute.provider_tree [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.413086] env[63418]: INFO nova.compute.manager [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Terminating instance [ 1072.569224] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e549fc9b-b555-4814-8299-2fc6ce1780ef tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.722s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.655146] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.794032] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 384b4ae7-ad55-4d44-bd10-3b82534c9703] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1072.915666] env[63418]: DEBUG nova.scheduler.client.report [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1072.922021] env[63418]: DEBUG nova.compute.manager [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1072.922021] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.922021] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64f1025-8767-4906-b71f-5932021fc166 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.933632] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.933956] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11e6548b-599a-4dde-85de-de7bfc70afbe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.942818] env[63418]: DEBUG oslo_vmware.api [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1072.942818] env[63418]: value = "task-1245572" [ 1072.942818] env[63418]: _type = "Task" [ 1072.942818] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.953778] env[63418]: DEBUG oslo_vmware.api [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.298755] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: dcc94934-d8d2-46d9-85fd-e17c11a7a4d7] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1073.421659] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.423924] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.769s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.447018] env[63418]: INFO nova.scheduler.client.report [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Deleted allocations for instance 1e6828a8-02ea-434f-b587-8d4c6854d944 [ 1073.458027] env[63418]: DEBUG oslo_vmware.api [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245572, 'name': PowerOffVM_Task, 'duration_secs': 0.164516} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.458321] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.458462] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.458716] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9973af62-f440-4c84-b8d0-ae1f287756fa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.519661] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.519661] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.519661] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Deleting the datastore file [datastore2] cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.519661] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbd9d33f-a755-47c5-b22c-5c564176d2b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.527757] env[63418]: DEBUG oslo_vmware.api [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1073.527757] env[63418]: value = "task-1245574" [ 1073.527757] env[63418]: _type = "Task" [ 1073.527757] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.539894] env[63418]: DEBUG oslo_vmware.api [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.800181] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 1f213c5b-de27-48d6-a1f8-fdf9b77359fd] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1073.927369] env[63418]: DEBUG nova.objects.instance [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'migration_context' on Instance uuid 5a8b347c-e9ec-4355-bcf7-b7d205da194c {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.956243] env[63418]: DEBUG oslo_concurrency.lockutils [None req-64eaa158-8282-4911-99d8-425c5ec404df tempest-MultipleCreateTestJSON-1959524581 tempest-MultipleCreateTestJSON-1959524581-project-member] Lock "1e6828a8-02ea-434f-b587-8d4c6854d944" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.975s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.976310] env[63418]: DEBUG oslo_concurrency.lockutils [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.976565] env[63418]: DEBUG oslo_concurrency.lockutils [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.976818] env[63418]: DEBUG nova.compute.manager [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1073.977985] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d3bb58-f606-4368-83ce-82733d2e8b28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.985561] env[63418]: DEBUG nova.compute.manager [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63418) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1073.986081] env[63418]: DEBUG nova.objects.instance [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'flavor' on Instance uuid 69e2b485-bb5d-4a90-8f78-66e6342b7928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.038431] env[63418]: DEBUG oslo_vmware.api [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21465} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.038810] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.039091] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.039359] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.039621] env[63418]: INFO nova.compute.manager [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1074.039889] env[63418]: DEBUG oslo.service.loopingcall [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.040116] env[63418]: DEBUG nova.compute.manager [-] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1074.040230] env[63418]: DEBUG nova.network.neutron [-] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1074.303951] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: e19a60cd-4ec7-48ad-9042-f19b75353364] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1074.529805] env[63418]: DEBUG nova.compute.manager [req-0af10319-5c74-439a-a376-77705747d3a2 req-13645149-5718-4a68-acc3-5d1fd6737b2d service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Received event network-vif-deleted-01df43f5-4069-46a0-b1e6-4baff131e04e {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1074.530060] env[63418]: INFO nova.compute.manager [req-0af10319-5c74-439a-a376-77705747d3a2 req-13645149-5718-4a68-acc3-5d1fd6737b2d service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Neutron deleted interface 01df43f5-4069-46a0-b1e6-4baff131e04e; detaching it from the instance and deleting it from the info cache [ 1074.530229] env[63418]: DEBUG nova.network.neutron [req-0af10319-5c74-439a-a376-77705747d3a2 req-13645149-5718-4a68-acc3-5d1fd6737b2d service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.595928] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5319ae01-70ac-45de-b7f8-efcf36eb6bde {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.604752] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ce993b-0cb5-4ba5-bf4f-553b95e00c05 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.635300] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117ebaf9-47a2-422f-bfbb-bfcda03f2fc0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.643781] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e130494-9160-4748-b560-beab07e205d7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.657815] env[63418]: DEBUG nova.compute.provider_tree [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.809008] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: e2dcb8b8-e778-4202-8808-6a8535e2f1b4] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1075.000036] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.000036] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7675131e-6430-4786-ae2d-16342c9ed4de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.003886] env[63418]: DEBUG oslo_vmware.api [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1075.003886] env[63418]: value = "task-1245575" [ 1075.003886] env[63418]: _type = "Task" [ 1075.003886] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.004123] env[63418]: DEBUG nova.network.neutron [-] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.015062] env[63418]: DEBUG oslo_vmware.api [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.033583] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d0b96b7-f621-44eb-879a-31aa97214c65 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.042541] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e01a60-f439-4d6c-8621-b7257986581d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.071137] env[63418]: DEBUG nova.compute.manager [req-0af10319-5c74-439a-a376-77705747d3a2 req-13645149-5718-4a68-acc3-5d1fd6737b2d service nova] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Detach interface failed, port_id=01df43f5-4069-46a0-b1e6-4baff131e04e, reason: Instance cbcbaf4f-ee24-4072-83ae-ffde59478928 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1075.160787] env[63418]: DEBUG nova.scheduler.client.report [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1075.313526] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 849552d4-0bb4-48ee-af7b-390183bd2189] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1075.509865] env[63418]: INFO nova.compute.manager [-] [instance: cbcbaf4f-ee24-4072-83ae-ffde59478928] Took 1.47 seconds to deallocate network for instance. [ 1075.521736] env[63418]: DEBUG oslo_vmware.api [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245575, 'name': PowerOffVM_Task, 'duration_secs': 0.280792} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.522227] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.522428] env[63418]: DEBUG nova.compute.manager [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1075.523217] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8764699f-01e1-4d08-b601-6f8a23908dec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.817390] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 6465fb5c-7bc9-4197-b3fb-bad2c000a5f8] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1076.024623] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.039839] env[63418]: DEBUG oslo_concurrency.lockutils [None req-867d1d28-89dd-40c1-813a-c113d7b380fc tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.063s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.170767] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.747s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.176507] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.152s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.176745] env[63418]: DEBUG nova.objects.instance [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'resources' on Instance uuid cbcbaf4f-ee24-4072-83ae-ffde59478928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.323991] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: f7554ae4-c7a7-4111-a830-10f9029dc074] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1076.788420] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce36515-fc70-482d-9ed1-ba00c82826bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.795940] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4c6e23-16a2-4ca7-b1be-07e3e13bd640 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.825458] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: fa4c0315-1ef5-4491-ab18-d49563b778fb] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1076.827618] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afab0c8-8cbd-4d1b-8695-d2a61a5ce8ea {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.835542] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cfca2c-977c-4203-8f8d-34d3e1b4eb59 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.850026] env[63418]: DEBUG nova.compute.provider_tree [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.859969] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.860212] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.860413] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "69e2b485-bb5d-4a90-8f78-66e6342b7928-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.860594] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.860759] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.862515] env[63418]: INFO nova.compute.manager [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Terminating instance [ 1077.330587] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7b0c70aa-e2bc-4131-97b4-4e53a378940a] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1077.352738] env[63418]: DEBUG nova.scheduler.client.report [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1077.365762] env[63418]: DEBUG nova.compute.manager [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1077.366012] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1077.367330] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9e84ea-d18e-492b-837d-0a46dd00e227 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.375362] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1077.375690] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71158335-4494-4a65-9625-4de38a297485 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.437764] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1077.437994] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1077.438183] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore1] 69e2b485-bb5d-4a90-8f78-66e6342b7928 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.438449] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a822e98e-1819-44b1-8733-d506337862e2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.445063] env[63418]: DEBUG oslo_vmware.api [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1077.445063] env[63418]: value = "task-1245577" [ 1077.445063] env[63418]: _type = "Task" [ 1077.445063] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.453565] env[63418]: DEBUG oslo_vmware.api [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.709473] env[63418]: INFO nova.compute.manager [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Swapping old allocation on dict_keys(['6ac9de28-4c58-4fc2-8a3d-711092e3c63c']) held by migration 1500a04c-32e3-42be-a525-4d03eac3038b for instance [ 1077.733956] env[63418]: DEBUG nova.scheduler.client.report [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Overwriting current allocation {'allocations': {'6ac9de28-4c58-4fc2-8a3d-711092e3c63c': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 144}}, 'project_id': '381605706a8b4a1e9cca8603c800f1ac', 'user_id': '1e9d38547e714839a808ae0c7e7e97d9', 'consumer_generation': 1} on consumer 5a8b347c-e9ec-4355-bcf7-b7d205da194c {{(pid=63418) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1077.809761] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.809959] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.810160] env[63418]: DEBUG nova.network.neutron [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1077.833178] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: be1c134c-f36e-4cc7-b4ef-8f30793fb4df] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1077.857429] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.873858] env[63418]: INFO nova.scheduler.client.report [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Deleted allocations for instance cbcbaf4f-ee24-4072-83ae-ffde59478928 [ 1077.954061] env[63418]: DEBUG oslo_vmware.api [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12546} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.954310] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1077.954499] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1077.954724] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1077.954937] env[63418]: INFO nova.compute.manager [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1077.955196] env[63418]: DEBUG oslo.service.loopingcall [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.955389] env[63418]: DEBUG nova.compute.manager [-] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1077.955482] env[63418]: DEBUG nova.network.neutron [-] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1078.227234] env[63418]: DEBUG nova.compute.manager [req-948d562a-c014-4311-b0b4-cc2ad2ecc998 req-002b6995-83ab-4b35-b850-517b4b5134bb service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Received event network-vif-deleted-d3da5ac7-6b95-4055-afd8-9e8f5a2d9657 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1078.227458] env[63418]: INFO nova.compute.manager [req-948d562a-c014-4311-b0b4-cc2ad2ecc998 req-002b6995-83ab-4b35-b850-517b4b5134bb service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Neutron deleted interface d3da5ac7-6b95-4055-afd8-9e8f5a2d9657; detaching it from the instance and deleting it from the info cache [ 1078.227622] env[63418]: DEBUG nova.network.neutron [req-948d562a-c014-4311-b0b4-cc2ad2ecc998 req-002b6995-83ab-4b35-b850-517b4b5134bb service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.336098] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: eaa13276-1fb1-47e2-ad1f-445bc9f4c98f] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1078.381180] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6435d4ee-a578-4bed-b149-e788ca0b1dc9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "cbcbaf4f-ee24-4072-83ae-ffde59478928" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.973s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.517997] env[63418]: DEBUG nova.network.neutron [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [{"id": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "address": "fa:16:3e:a1:9c:70", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eb8c1f9-ee", "ovs_interfaceid": "6eb8c1f9-eec0-4763-8033-34a4efda65c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.702595] env[63418]: DEBUG nova.network.neutron [-] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.731342] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f01617cc-5257-4ade-90c0-c6b026c4fdc7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.741504] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3183ae2-9257-4eff-94bb-5ebccca8a743 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.768279] env[63418]: DEBUG nova.compute.manager [req-948d562a-c014-4311-b0b4-cc2ad2ecc998 req-002b6995-83ab-4b35-b850-517b4b5134bb service nova] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Detach interface failed, port_id=d3da5ac7-6b95-4055-afd8-9e8f5a2d9657, reason: Instance 69e2b485-bb5d-4a90-8f78-66e6342b7928 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1078.840022] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7ebfdcb9-79f2-444d-b63a-8aaad9bc3a00] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1079.021019] env[63418]: DEBUG oslo_concurrency.lockutils [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-5a8b347c-e9ec-4355-bcf7-b7d205da194c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.022150] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4244dbed-41ed-432a-afa0-743d2a88bdb8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.032396] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee81de1-0554-473f-959e-35fa2f6d6258 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.205108] env[63418]: INFO nova.compute.manager [-] [instance: 69e2b485-bb5d-4a90-8f78-66e6342b7928] Took 1.25 seconds to deallocate network for instance. [ 1079.343376] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 27290558-6c58-414d-bd53-c4a686890721] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1079.712449] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.712902] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.713284] env[63418]: DEBUG nova.objects.instance [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid 69e2b485-bb5d-4a90-8f78-66e6342b7928 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.848310] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 0de234c4-1b8d-4b9f-8d32-cbf2b17f2d48] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1080.137291] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.137610] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-184360f2-b2de-40a0-8778-4ff86043002d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.147053] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1080.147053] env[63418]: value = "task-1245578" [ 1080.147053] env[63418]: _type = "Task" [ 1080.147053] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.153684] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.309890] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431f6aa9-1912-400b-9ef1-5a08df67c668 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.318373] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757c2f26-94eb-4ba7-8a38-bb56eb701ee1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.348538] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31cb646-ee00-4a78-a174-6e867e3841a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.351054] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 18327b62-d022-45e5-9ab0-71791491e61f] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1080.357524] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab881f2d-ae35-47d3-aeac-7477d92c7b78 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.372171] env[63418]: DEBUG nova.compute.provider_tree [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.654862] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245578, 'name': PowerOffVM_Task, 'duration_secs': 0.192035} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.655151] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.655830] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.656085] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.656578] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.656578] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.656717] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.656753] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.656987] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.657147] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.657322] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.657486] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.657660] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.662698] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eef1098b-5b33-45fa-ba7f-9335297a1188 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.678385] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1080.678385] env[63418]: value = "task-1245579" [ 1080.678385] env[63418]: _type = "Task" [ 1080.678385] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.686326] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.854234] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 1a57a23c-3b9d-4b6b-ae0b-4b12340e96e5] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1080.875595] env[63418]: DEBUG nova.scheduler.client.report [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1080.948953] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.949197] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.188346] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245579, 'name': ReconfigVM_Task, 'duration_secs': 0.151171} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.189164] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de489bc-d814-46ce-bd08-fb729ee8b4da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.209822] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.210071] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.210242] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.210433] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.210583] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.210734] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.210940] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.211118] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.211294] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.211459] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.211637] env[63418]: DEBUG nova.virt.hardware [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.212411] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ae29c3-b60e-475b-bc85-cb5a3ef8ed4f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.217626] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1081.217626] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]527a4bb9-1eae-bf06-a5da-81b8b237cd6a" [ 1081.217626] env[63418]: _type = "Task" [ 1081.217626] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.225826] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527a4bb9-1eae-bf06-a5da-81b8b237cd6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.357751] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 1f898bd4-0b80-48c1-9e7c-8948f4b19c0c] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1081.380565] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.668s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.406237] env[63418]: INFO nova.scheduler.client.report [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance 69e2b485-bb5d-4a90-8f78-66e6342b7928 [ 1081.451322] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1081.727411] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]527a4bb9-1eae-bf06-a5da-81b8b237cd6a, 'name': SearchDatastore_Task, 'duration_secs': 0.00736} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.732650] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1081.732921] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af8e2e81-8ce2-4644-866b-ca01b3a9c0de {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.752788] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1081.752788] env[63418]: value = "task-1245580" [ 1081.752788] env[63418]: _type = "Task" [ 1081.752788] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.761264] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245580, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.861736] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: e3dd6d5e-2b80-4724-9826-6ebfb8181c3b] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1081.912979] env[63418]: DEBUG oslo_concurrency.lockutils [None req-34511297-9ba0-4da0-9718-3042fa1622d7 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "69e2b485-bb5d-4a90-8f78-66e6342b7928" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.053s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.970410] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.970687] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.972294] env[63418]: INFO nova.compute.claims [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.262724] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245580, 'name': ReconfigVM_Task, 'duration_secs': 0.223755} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.263274] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1082.264051] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ac4b78-da0d-4873-ac3c-126ecab164f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.288186] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.288428] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0128ac16-5fc0-4d52-a32a-4b17318701e5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.305257] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1082.305257] env[63418]: value = "task-1245581" [ 1082.305257] env[63418]: _type = "Task" [ 1082.305257] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.312253] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.365150] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: f632d71d-498f-4914-9895-8f37187a295f] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1082.625170] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "bf091892-fefb-49dd-9416-708a06b35798" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.625451] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "bf091892-fefb-49dd-9416-708a06b35798" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.625671] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "bf091892-fefb-49dd-9416-708a06b35798-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.625873] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "bf091892-fefb-49dd-9416-708a06b35798-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.626107] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "bf091892-fefb-49dd-9416-708a06b35798-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.628338] env[63418]: INFO nova.compute.manager [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Terminating instance [ 1082.815062] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245581, 'name': ReconfigVM_Task, 'duration_secs': 0.267458} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.815356] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c/5a8b347c-e9ec-4355-bcf7-b7d205da194c.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.816198] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e55cca8-5e57-4a90-b4e9-f84fd42dd5d4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.835776] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8eeea33-967b-48cd-9a5d-a43fbd830d81 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.855917] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7458b6d-d794-4f79-8c28-90021979bdb9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.875386] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 0c151b0c-c383-4421-b506-b7afa95e2072] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1082.877747] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad91f390-b18d-44ad-8377-da1b03bbf90a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.884764] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.885012] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d89763f-3ac2-4ee5-8d93-ad4d8f8fd2fc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.890995] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1082.890995] env[63418]: value = "task-1245582" [ 1082.890995] env[63418]: _type = "Task" [ 1082.890995] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.898297] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.068467] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23742372-ff9d-47b2-829f-16c7f1a0b00f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.076168] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e709acee-e4ad-4e30-8144-8b24c2a14fe6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.107725] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b922f4-6813-4b46-829e-b163b6b69d7d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.114885] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74022db5-5d88-440e-9d84-2c42a2e7a0b3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.128049] env[63418]: DEBUG nova.compute.provider_tree [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.132019] env[63418]: DEBUG nova.compute.manager [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1083.132249] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1083.133029] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95306a24-5b62-4586-8fdc-a0b956084285 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.140440] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.140670] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e84d9af-2a79-4bfd-9429-265f837fcebd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.145955] env[63418]: DEBUG oslo_vmware.api [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1083.145955] env[63418]: value = "task-1245583" [ 1083.145955] env[63418]: _type = "Task" [ 1083.145955] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.153655] env[63418]: DEBUG oslo_vmware.api [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.381530] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7615244b-fa19-4dbb-a851-604bb6fc2725] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1083.399896] env[63418]: DEBUG oslo_vmware.api [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245582, 'name': PowerOnVM_Task, 'duration_secs': 0.372195} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.400213] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.631518] env[63418]: DEBUG nova.scheduler.client.report [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1083.655844] env[63418]: DEBUG oslo_vmware.api [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245583, 'name': PowerOffVM_Task, 'duration_secs': 0.200146} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.656171] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1083.656353] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1083.656591] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebbc1020-3fa9-4f23-b513-9eb5dc3de002 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.715377] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1083.715579] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1083.715767] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleting the datastore file [datastore2] bf091892-fefb-49dd-9416-708a06b35798 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.716085] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96438537-0a10-4b7d-a63f-e5d4a00bf262 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.723210] env[63418]: DEBUG oslo_vmware.api [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for the task: (returnval){ [ 1083.723210] env[63418]: value = "task-1245585" [ 1083.723210] env[63418]: _type = "Task" [ 1083.723210] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.730702] env[63418]: DEBUG oslo_vmware.api [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.885399] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 3e7b9a60-9e4b-46e3-adf9-7337dc7d2171] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1084.136881] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.137431] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1084.233600] env[63418]: DEBUG oslo_vmware.api [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Task: {'id': task-1245585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.304533} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.233950] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.234206] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1084.234492] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1084.234816] env[63418]: INFO nova.compute.manager [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] [instance: bf091892-fefb-49dd-9416-708a06b35798] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1084.235190] env[63418]: DEBUG oslo.service.loopingcall [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.235480] env[63418]: DEBUG nova.compute.manager [-] [instance: bf091892-fefb-49dd-9416-708a06b35798] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1084.235618] env[63418]: DEBUG nova.network.neutron [-] [instance: bf091892-fefb-49dd-9416-708a06b35798] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1084.389772] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: c4cb1770-b3c7-4f6b-ba11-85310f2175a0] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1084.455592] env[63418]: INFO nova.compute.manager [None req-eef79194-78f0-462a-a113-79b640b55980 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance to original state: 'active' [ 1084.518093] env[63418]: DEBUG nova.compute.manager [req-1d7b87b2-84e7-49be-a9c5-a4edff66ca01 req-3e6e70bf-d15d-407a-9f27-4800e72a9c24 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Received event network-vif-deleted-b1cdc19c-4c4d-4568-ade7-c0b97713632b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1084.518374] env[63418]: INFO nova.compute.manager [req-1d7b87b2-84e7-49be-a9c5-a4edff66ca01 req-3e6e70bf-d15d-407a-9f27-4800e72a9c24 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Neutron deleted interface b1cdc19c-4c4d-4568-ade7-c0b97713632b; detaching it from the instance and deleting it from the info cache [ 1084.518566] env[63418]: DEBUG nova.network.neutron [req-1d7b87b2-84e7-49be-a9c5-a4edff66ca01 req-3e6e70bf-d15d-407a-9f27-4800e72a9c24 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.642793] env[63418]: DEBUG nova.compute.utils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1084.644513] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1084.644825] env[63418]: DEBUG nova.network.neutron [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1084.692803] env[63418]: DEBUG nova.policy [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94b1aab9fe174e8caa3a809e1b642131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aff8e2c3d27c4277b2b8f130c293c013', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1084.893721] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d76a008c-9bd9-420b-873d-4f7d7f25b8ca] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1084.980135] env[63418]: DEBUG nova.network.neutron [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Successfully created port: 22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1084.994535] env[63418]: DEBUG nova.network.neutron [-] [instance: bf091892-fefb-49dd-9416-708a06b35798] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.020942] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6559d2ab-6926-4109-be7b-d36cb27ba867 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.031317] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2278c069-85b8-4f69-b70c-9bf12ff77537 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.062805] env[63418]: DEBUG nova.compute.manager [req-1d7b87b2-84e7-49be-a9c5-a4edff66ca01 req-3e6e70bf-d15d-407a-9f27-4800e72a9c24 service nova] [instance: bf091892-fefb-49dd-9416-708a06b35798] Detach interface failed, port_id=b1cdc19c-4c4d-4568-ade7-c0b97713632b, reason: Instance bf091892-fefb-49dd-9416-708a06b35798 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1085.148215] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1085.397717] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: c0fd1999-346d-4c12-9b8c-ab7e21ec4227] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1085.497462] env[63418]: INFO nova.compute.manager [-] [instance: bf091892-fefb-49dd-9416-708a06b35798] Took 1.26 seconds to deallocate network for instance. [ 1085.901067] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 17a055e4-13da-4984-ae64-af1ae9d652f6] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1086.004130] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.004441] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.004665] env[63418]: DEBUG nova.objects.instance [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lazy-loading 'resources' on Instance uuid bf091892-fefb-49dd-9416-708a06b35798 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.144033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.144033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.144033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.144033] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.144371] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.146848] env[63418]: INFO nova.compute.manager [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Terminating instance [ 1086.157840] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1086.183236] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.183477] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.183637] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.183818] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.183999] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.184161] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.184393] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.184562] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.184729] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.184892] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.185080] env[63418]: DEBUG nova.virt.hardware [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.185925] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf9eb33-fb35-44bd-ae3a-5e0581bdd4ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.193978] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190f9aff-c388-4853-a7c2-028b4d203676 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.404332] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 6b81dbe2-aa9e-4561-962b-2af167234b90] Instance has had 0 of 5 cleanup attempts {{(pid=63418) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11591}} [ 1086.440238] env[63418]: DEBUG nova.network.neutron [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Successfully updated port: 22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1086.549658] env[63418]: DEBUG nova.compute.manager [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Received event network-vif-plugged-22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1086.549867] env[63418]: DEBUG oslo_concurrency.lockutils [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.550096] env[63418]: DEBUG oslo_concurrency.lockutils [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.550249] env[63418]: DEBUG oslo_concurrency.lockutils [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.550420] env[63418]: DEBUG nova.compute.manager [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] No waiting events found dispatching network-vif-plugged-22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1086.550589] env[63418]: WARNING nova.compute.manager [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Received unexpected event network-vif-plugged-22836017-f955-4526-9615-0c23c172ddf2 for instance with vm_state building and task_state spawning. [ 1086.550751] env[63418]: DEBUG nova.compute.manager [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Received event network-changed-22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1086.550907] env[63418]: DEBUG nova.compute.manager [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Refreshing instance network info cache due to event network-changed-22836017-f955-4526-9615-0c23c172ddf2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1086.551111] env[63418]: DEBUG oslo_concurrency.lockutils [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] Acquiring lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.551261] env[63418]: DEBUG oslo_concurrency.lockutils [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] Acquired lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.551480] env[63418]: DEBUG nova.network.neutron [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Refreshing network info cache for port 22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1086.600058] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612d9416-ca7c-41a9-acb3-1a3c2489ccac {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.608013] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b13abb-5a53-41f2-a40f-84917d917995 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.640852] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84b4efd-7e1e-41ab-8ec3-f58251e76d8c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.649205] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82382da-7a68-4277-af13-0994bbdac556 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.653485] env[63418]: DEBUG nova.compute.manager [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1086.653704] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.653923] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-322dbdf9-a471-454f-b226-713dc4eba3c4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.665155] env[63418]: DEBUG nova.compute.provider_tree [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.667417] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1086.667417] env[63418]: value = "task-1245586" [ 1086.667417] env[63418]: _type = "Task" [ 1086.667417] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.676673] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245586, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.907960] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.942652] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.085767] env[63418]: DEBUG nova.network.neutron [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1087.164182] env[63418]: DEBUG nova.network.neutron [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.168825] env[63418]: DEBUG nova.scheduler.client.report [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1087.181734] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245586, 'name': PowerOffVM_Task, 'duration_secs': 0.222933} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.181993] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.182208] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.182405] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268541', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'name': 'volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '5a8b347c-e9ec-4355-bcf7-b7d205da194c', 'attached_at': '2024-10-10T13:47:21.000000', 'detached_at': '', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'serial': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.183177] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca94dc0-4462-44f8-836c-468b7557b739 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.204173] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4051a6db-62e2-4a6d-8b5d-f3a1a3472f18 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.212845] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b99c5ca-ff0a-4a02-be6e-5e6b139b3d89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.232721] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eded456-09b4-430b-bc32-ee7d85544955 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.248118] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] The volume has not been displaced from its original location: [datastore1] volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5/volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1087.253204] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1087.253475] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62177c1a-7de8-4ddc-90ea-3fdde4d584b0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.271630] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1087.271630] env[63418]: value = "task-1245587" [ 1087.271630] env[63418]: _type = "Task" [ 1087.271630] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.279912] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245587, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.667639] env[63418]: DEBUG oslo_concurrency.lockutils [req-47327697-477f-4769-80fd-f368fdc7f9cc req-af1a0036-e722-47e0-81c4-0cf7afaf771e service nova] Releasing lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.668069] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.668240] env[63418]: DEBUG nova.network.neutron [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.677060] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.700810] env[63418]: INFO nova.scheduler.client.report [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Deleted allocations for instance bf091892-fefb-49dd-9416-708a06b35798 [ 1087.782074] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245587, 'name': ReconfigVM_Task, 'duration_secs': 0.212982} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.782379] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1087.786966] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abec7a0a-83b2-4112-9f6d-628791aad0cf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.801482] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1087.801482] env[63418]: value = "task-1245588" [ 1087.801482] env[63418]: _type = "Task" [ 1087.801482] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.809122] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.974517] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.974829] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.975279] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.975279] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.975418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.977785] env[63418]: INFO nova.compute.manager [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Terminating instance [ 1088.199736] env[63418]: DEBUG nova.network.neutron [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1088.208054] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7a5aeb31-c1b7-44a5-8355-a4be51a38fd1 tempest-ServersTestJSON-1949460379 tempest-ServersTestJSON-1949460379-project-member] Lock "bf091892-fefb-49dd-9416-708a06b35798" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.582s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.312660] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245588, 'name': ReconfigVM_Task, 'duration_secs': 0.154185} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.312966] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268541', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'name': 'volume-6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '5a8b347c-e9ec-4355-bcf7-b7d205da194c', 'attached_at': '2024-10-10T13:47:21.000000', 'detached_at': '', 'volume_id': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5', 'serial': '6e5bcd2d-214c-4a04-8d2a-28ee190996b5'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1088.313232] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.313971] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416635bc-d223-4dff-9e9d-e0b9d5ad073a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.320172] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1088.320958] env[63418]: DEBUG nova.network.neutron [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating instance_info_cache with network_info: [{"id": "22836017-f955-4526-9615-0c23c172ddf2", "address": "fa:16:3e:5a:2b:21", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22836017-f9", "ovs_interfaceid": "22836017-f955-4526-9615-0c23c172ddf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.322039] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8abdbf5-08a5-441a-a5da-7dc96e105917 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.392117] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1088.392521] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1088.392735] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleting the datastore file [datastore1] 5a8b347c-e9ec-4355-bcf7-b7d205da194c {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.393020] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3afb9c03-ad8b-4539-a890-fdb0282f0271 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.399767] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1088.399767] env[63418]: value = "task-1245590" [ 1088.399767] env[63418]: _type = "Task" [ 1088.399767] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.407933] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1088.409358] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.482424] env[63418]: DEBUG nova.compute.manager [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1088.482691] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.483822] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8941ff-9c15-459e-995b-b7f960197b15 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.492654] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.492919] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5509a067-2264-45de-a5ff-41e71e4315e4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.499290] env[63418]: DEBUG oslo_vmware.api [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1088.499290] env[63418]: value = "task-1245591" [ 1088.499290] env[63418]: _type = "Task" [ 1088.499290] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.506897] env[63418]: DEBUG oslo_vmware.api [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.824355] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.824697] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Instance network_info: |[{"id": "22836017-f955-4526-9615-0c23c172ddf2", "address": "fa:16:3e:5a:2b:21", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22836017-f9", "ovs_interfaceid": "22836017-f955-4526-9615-0c23c172ddf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1088.825161] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:2b:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22836017-f955-4526-9615-0c23c172ddf2', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.832778] env[63418]: DEBUG oslo.service.loopingcall [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.832983] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.833230] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31e6f2f2-2ce9-4ed7-95bb-2c67b33d819e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.852643] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.852643] env[63418]: value = "task-1245592" [ 1088.852643] env[63418]: _type = "Task" [ 1088.852643] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.859788] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245592, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.908600] env[63418]: DEBUG oslo_vmware.api [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245590, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156964} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.908900] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.909201] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1088.909316] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1088.909516] env[63418]: INFO nova.compute.manager [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1088.909755] env[63418]: DEBUG oslo.service.loopingcall [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.910094] env[63418]: DEBUG nova.compute.manager [-] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1088.910094] env[63418]: DEBUG nova.network.neutron [-] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1088.914457] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1088.914538] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 1088.914622] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Rebuilding the list of instances to heal {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10282}} [ 1089.009041] env[63418]: DEBUG oslo_vmware.api [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245591, 'name': PowerOffVM_Task, 'duration_secs': 0.207513} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.009409] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.009595] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.009852] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a40392b-cc0e-4af9-bec8-6d93d18ca548 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.073161] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.073401] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deleting contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.073592] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleting the datastore file [datastore1] d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.073906] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-352d2413-c3f5-47ca-9fbf-6f2b14e05f28 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.080409] env[63418]: DEBUG oslo_vmware.api [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for the task: (returnval){ [ 1089.080409] env[63418]: value = "task-1245594" [ 1089.080409] env[63418]: _type = "Task" [ 1089.080409] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.089037] env[63418]: DEBUG oslo_vmware.api [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.360423] env[63418]: DEBUG nova.compute.manager [req-449ef6de-fe33-4fa7-a98d-cd1947834088 req-c440f0cb-2fa2-4ea8-aec7-86f0099b0780 service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Received event network-vif-deleted-6eb8c1f9-eec0-4763-8033-34a4efda65c1 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1089.360423] env[63418]: INFO nova.compute.manager [req-449ef6de-fe33-4fa7-a98d-cd1947834088 req-c440f0cb-2fa2-4ea8-aec7-86f0099b0780 service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Neutron deleted interface 6eb8c1f9-eec0-4763-8033-34a4efda65c1; detaching it from the instance and deleting it from the info cache [ 1089.360423] env[63418]: DEBUG nova.network.neutron [req-449ef6de-fe33-4fa7-a98d-cd1947834088 req-c440f0cb-2fa2-4ea8-aec7-86f0099b0780 service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.367776] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245592, 'name': CreateVM_Task, 'duration_secs': 0.332814} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.367977] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1089.368599] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.368793] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.369119] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1089.369369] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77302792-058e-429d-b0a9-1982173ac966 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.374977] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1089.374977] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52205f6d-e457-f02b-4fed-75369102ae41" [ 1089.374977] env[63418]: _type = "Task" [ 1089.374977] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.382780] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52205f6d-e457-f02b-4fed-75369102ae41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.419389] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Skipping network cache update for instance because it is being deleted. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10295}} [ 1089.419539] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Skipping network cache update for instance because it is being deleted. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10295}} [ 1089.419665] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 1089.445711] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.445828] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquired lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.445918] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Forcefully refreshing network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1089.446097] env[63418]: DEBUG nova.objects.instance [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lazy-loading 'info_cache' on Instance uuid 98f51fcd-951c-4c16-bc20-efc62ef359f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.598101] env[63418]: DEBUG oslo_vmware.api [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Task: {'id': task-1245594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140386} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.600877] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.601551] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deleted contents of the VM from datastore datastore1 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.602677] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.602927] env[63418]: INFO nova.compute.manager [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1089.603237] env[63418]: DEBUG oslo.service.loopingcall [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.603482] env[63418]: DEBUG nova.compute.manager [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1089.603581] env[63418]: DEBUG nova.network.neutron [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.836466] env[63418]: DEBUG nova.network.neutron [-] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.863911] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0ffb0df-f1db-43ff-acda-5177b49ae1b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.875473] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954dce54-bb51-4d3f-94f6-3a592af9bab7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.914943] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52205f6d-e457-f02b-4fed-75369102ae41, 'name': SearchDatastore_Task, 'duration_secs': 0.008456} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.924433] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.925093] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.925093] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.925242] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.925326] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.925741] env[63418]: DEBUG nova.compute.manager [req-449ef6de-fe33-4fa7-a98d-cd1947834088 req-c440f0cb-2fa2-4ea8-aec7-86f0099b0780 service nova] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Detach interface failed, port_id=6eb8c1f9-eec0-4763-8033-34a4efda65c1, reason: Instance 5a8b347c-e9ec-4355-bcf7-b7d205da194c could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1089.926193] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54201c18-7bc8-40d8-97f0-7bf86233ced1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.935839] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.936047] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1089.936781] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37e58c75-6e61-44ca-8aab-4740fd763d18 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.942032] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1089.942032] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]529d5c07-2fd7-aa73-0fa3-19fcfac8a2ef" [ 1089.942032] env[63418]: _type = "Task" [ 1089.942032] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.952404] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529d5c07-2fd7-aa73-0fa3-19fcfac8a2ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.344021] env[63418]: INFO nova.compute.manager [-] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Took 1.43 seconds to deallocate network for instance. [ 1090.454926] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]529d5c07-2fd7-aa73-0fa3-19fcfac8a2ef, 'name': SearchDatastore_Task, 'duration_secs': 0.010132} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.456301] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5068f921-1bb3-4701-9471-facb49f8f7ee {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.462801] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1090.462801] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c9a1c9-a52a-4fdf-617f-bd7638b8a028" [ 1090.462801] env[63418]: _type = "Task" [ 1090.462801] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.473020] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c9a1c9-a52a-4fdf-617f-bd7638b8a028, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.547919] env[63418]: DEBUG nova.network.neutron [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.771997] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "fe38702e-953d-4748-a523-401e798b2271" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.772321] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "fe38702e-953d-4748-a523-401e798b2271" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.772544] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "fe38702e-953d-4748-a523-401e798b2271-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.772734] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "fe38702e-953d-4748-a523-401e798b2271-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.772909] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "fe38702e-953d-4748-a523-401e798b2271-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.775149] env[63418]: INFO nova.compute.manager [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Terminating instance [ 1090.893680] env[63418]: INFO nova.compute.manager [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 5a8b347c-e9ec-4355-bcf7-b7d205da194c] Took 0.55 seconds to detach 1 volumes for instance. [ 1090.982393] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52c9a1c9-a52a-4fdf-617f-bd7638b8a028, 'name': SearchDatastore_Task, 'duration_secs': 0.008606} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.982769] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.983110] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 7d532ab0-bf4c-402c-ac81-892930d13d6b/7d532ab0-bf4c-402c-ac81-892930d13d6b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.983409] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f834e98-bd64-4165-8b2a-4045815454d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.994824] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1090.994824] env[63418]: value = "task-1245595" [ 1090.994824] env[63418]: _type = "Task" [ 1090.994824] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.002637] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.049431] env[63418]: INFO nova.compute.manager [-] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Took 1.45 seconds to deallocate network for instance. [ 1091.189808] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating instance_info_cache with network_info: [{"id": "a6337c68-709c-4b53-b00a-b0e794cd45db", "address": "fa:16:3e:78:ae:27", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6337c68-70", "ovs_interfaceid": "a6337c68-709c-4b53-b00a-b0e794cd45db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.282428] env[63418]: DEBUG nova.compute.manager [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1091.282777] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.283718] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e506d60-8200-41cd-a705-c84bd9546138 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.292420] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.292708] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0ee0f5e-f03b-4b39-ada7-a2496265a1a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.299695] env[63418]: DEBUG oslo_vmware.api [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1091.299695] env[63418]: value = "task-1245596" [ 1091.299695] env[63418]: _type = "Task" [ 1091.299695] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.308685] env[63418]: DEBUG oslo_vmware.api [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.389786] env[63418]: DEBUG nova.compute.manager [req-7e7cd5c3-6f15-4cc4-b9a6-6268e4ef77d1 req-6618545d-b495-48ca-a5fb-6caebdd1e455 service nova] [instance: d52db46b-2461-4bd0-be57-d414250aac7e] Received event network-vif-deleted-9cc10d6a-03cf-43e3-84f7-9160f8626ad4 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1091.400987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.401287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.401567] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.428906] env[63418]: INFO nova.scheduler.client.report [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted allocations for instance 5a8b347c-e9ec-4355-bcf7-b7d205da194c [ 1091.506437] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245595, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.41023} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.506728] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 7d532ab0-bf4c-402c-ac81-892930d13d6b/7d532ab0-bf4c-402c-ac81-892930d13d6b.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.506947] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.507239] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef4fb8a6-d007-4f12-8cb1-4deeaf02d848 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.513922] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1091.513922] env[63418]: value = "task-1245597" [ 1091.513922] env[63418]: _type = "Task" [ 1091.513922] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.522636] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245597, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.555457] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.555802] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.556047] env[63418]: DEBUG nova.objects.instance [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lazy-loading 'resources' on Instance uuid d52db46b-2461-4bd0-be57-d414250aac7e {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.692637] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Releasing lock "refresh_cache-98f51fcd-951c-4c16-bc20-efc62ef359f5" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.692803] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updated the network info_cache for instance {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10349}} [ 1091.693030] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.693202] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.693557] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.693557] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.693661] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.693765] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 1091.693906] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.810947] env[63418]: DEBUG oslo_vmware.api [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245596, 'name': PowerOffVM_Task, 'duration_secs': 0.244536} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.811233] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.811429] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.811690] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13b5454d-c616-4029-b322-942028a9002a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.874666] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.874904] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.875105] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleting the datastore file [datastore2] fe38702e-953d-4748-a523-401e798b2271 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.875391] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c2720b3-b460-47a9-91a0-62df73177a4d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.882314] env[63418]: DEBUG oslo_vmware.api [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1091.882314] env[63418]: value = "task-1245599" [ 1091.882314] env[63418]: _type = "Task" [ 1091.882314] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.890241] env[63418]: DEBUG oslo_vmware.api [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.937370] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4bad02ef-29ad-4e8f-89e2-3f28d9088c53 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "5a8b347c-e9ec-4355-bcf7-b7d205da194c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.794s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.025088] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245597, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07223} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.025401] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1092.026285] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f944843a-44b5-48c6-a7d5-1b5e277b844e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.048770] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 7d532ab0-bf4c-402c-ac81-892930d13d6b/7d532ab0-bf4c-402c-ac81-892930d13d6b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.049076] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cf6dc07-47e9-4277-b71b-3059ca7aad45 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.070218] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1092.070218] env[63418]: value = "task-1245600" [ 1092.070218] env[63418]: _type = "Task" [ 1092.070218] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.078461] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.121081] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3ac274-1e84-419e-972e-32e0e7c6012d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.128148] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f2f10b-c834-43ea-97e6-d9ef7b4b7c6b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.157235] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96b4f58-9c02-4862-b8a0-c2d8b7d81e34 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.163987] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1dcd44d-29c6-44c7-9019-56fd0f983755 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.178085] env[63418]: DEBUG nova.compute.provider_tree [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.196924] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.391988] env[63418]: DEBUG oslo_vmware.api [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142441} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.392324] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.392456] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.392635] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.392811] env[63418]: INFO nova.compute.manager [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: fe38702e-953d-4748-a523-401e798b2271] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1092.393076] env[63418]: DEBUG oslo.service.loopingcall [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.393277] env[63418]: DEBUG nova.compute.manager [-] [instance: fe38702e-953d-4748-a523-401e798b2271] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1092.393370] env[63418]: DEBUG nova.network.neutron [-] [instance: fe38702e-953d-4748-a523-401e798b2271] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1092.580059] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245600, 'name': ReconfigVM_Task, 'duration_secs': 0.274733} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.580380] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 7d532ab0-bf4c-402c-ac81-892930d13d6b/7d532ab0-bf4c-402c-ac81-892930d13d6b.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.581054] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2bc8594-6b4a-42b0-840b-560bf4185d9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.587556] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1092.587556] env[63418]: value = "task-1245601" [ 1092.587556] env[63418]: _type = "Task" [ 1092.587556] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.598130] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245601, 'name': Rename_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.672712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.672962] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.680696] env[63418]: DEBUG nova.scheduler.client.report [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1093.098139] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245601, 'name': Rename_Task, 'duration_secs': 0.135176} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.098644] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1093.101074] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-198bbe2d-4757-456a-932f-d2ddfbb58208 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.105435] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1093.105435] env[63418]: value = "task-1245602" [ 1093.105435] env[63418]: _type = "Task" [ 1093.105435] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.113688] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.177937] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1093.185552] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.630s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.187875] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.991s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.188136] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.188566] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1093.189596] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf6dbfc-f42a-4877-ae1a-f792ca624c67 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.198071] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94217fe2-0374-464e-b6ee-ca4b0c061ca6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.214021] env[63418]: INFO nova.scheduler.client.report [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Deleted allocations for instance d52db46b-2461-4bd0-be57-d414250aac7e [ 1093.215782] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae7d0bc-6ce4-4bb7-aa0a-3432afa81bc8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.226564] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0ab7c8-d773-4079-84e5-67ab4ee42e21 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.257786] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180409MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1093.258031] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.258291] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.328623] env[63418]: DEBUG nova.network.neutron [-] [instance: fe38702e-953d-4748-a523-401e798b2271] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.415555] env[63418]: DEBUG nova.compute.manager [req-b19155d7-37b8-4615-8b55-c58a18867015 req-63efcc43-6338-4966-902a-8f4517de0749 service nova] [instance: fe38702e-953d-4748-a523-401e798b2271] Received event network-vif-deleted-9e929283-d11e-4069-ab19-e7195f2cfd1b {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1093.615715] env[63418]: DEBUG oslo_vmware.api [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245602, 'name': PowerOnVM_Task, 'duration_secs': 0.448774} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.615988] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1093.616212] env[63418]: INFO nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Took 7.46 seconds to spawn the instance on the hypervisor. [ 1093.616419] env[63418]: DEBUG nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1093.617261] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15ce105-4870-4be1-9d72-91b3b02f15e3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.696655] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.724985] env[63418]: DEBUG oslo_concurrency.lockutils [None req-acd694d7-eb3c-4206-872b-ab4fdddd26a9 tempest-AttachVolumeShelveTestJSON-1541482774 tempest-AttachVolumeShelveTestJSON-1541482774-project-member] Lock "d52db46b-2461-4bd0-be57-d414250aac7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.750s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.831211] env[63418]: INFO nova.compute.manager [-] [instance: fe38702e-953d-4748-a523-401e798b2271] Took 1.44 seconds to deallocate network for instance. [ 1094.133190] env[63418]: INFO nova.compute.manager [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Took 12.18 seconds to build instance. [ 1094.280083] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 98f51fcd-951c-4c16-bc20-efc62ef359f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.280246] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance fe38702e-953d-4748-a523-401e798b2271 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.280397] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7d532ab0-bf4c-402c-ac81-892930d13d6b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.337718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.635847] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2c5a923f-e3df-4aae-8448-7e46c4d7cafa tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.686s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.657432] env[63418]: DEBUG nova.compute.manager [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Received event network-changed-22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1094.657615] env[63418]: DEBUG nova.compute.manager [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Refreshing instance network info cache due to event network-changed-22836017-f955-4526-9615-0c23c172ddf2. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1094.657832] env[63418]: DEBUG oslo_concurrency.lockutils [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] Acquiring lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.658078] env[63418]: DEBUG oslo_concurrency.lockutils [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] Acquired lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.658159] env[63418]: DEBUG nova.network.neutron [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Refreshing network info cache for port 22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1094.783495] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 95240ada-686c-484b-9c68-bc9a1f3d3e70 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1094.783720] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1094.783869] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1094.850950] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c235c9-5dd3-41a6-91ab-ac57ba902ebf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.858949] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c948c098-ae26-4f78-ab04-abcfbdbdd6c9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.888013] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bca54fb-5d64-4478-b443-ef44361b242a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.895547] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e27ec7-92f2-4cd8-8272-07fcf31d2492 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.908618] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.363762] env[63418]: DEBUG nova.network.neutron [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updated VIF entry in instance network info cache for port 22836017-f955-4526-9615-0c23c172ddf2. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.364183] env[63418]: DEBUG nova.network.neutron [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating instance_info_cache with network_info: [{"id": "22836017-f955-4526-9615-0c23c172ddf2", "address": "fa:16:3e:5a:2b:21", "network": {"id": "bb2e7cf8-f935-48a3-9ef3-922f104f758b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-605586067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aff8e2c3d27c4277b2b8f130c293c013", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22836017-f9", "ovs_interfaceid": "22836017-f955-4526-9615-0c23c172ddf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.411755] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1095.867397] env[63418]: DEBUG oslo_concurrency.lockutils [req-5347842f-7b85-4ccd-a234-0178f8fcce20 req-2085f6c4-2c1d-4a5f-a1a3-d51d11fecdf4 service nova] Releasing lock "refresh_cache-7d532ab0-bf4c-402c-ac81-892930d13d6b" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.916398] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1095.916620] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.658s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.916901] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.220s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.918509] env[63418]: INFO nova.compute.claims [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.982226] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6c7875-8394-4b54-88d4-4dd556501edd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.990158] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e791816b-da97-4e08-a553-2d00c922e1f3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.019213] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb1dc07-014c-4577-914a-0a1517296fa2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.026304] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8390ea24-fff4-498e-9a07-f42e064b35f8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.038945] env[63418]: DEBUG nova.compute.provider_tree [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.542228] env[63418]: DEBUG nova.scheduler.client.report [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1098.047127] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.047826] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1098.051533] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.714s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.052295] env[63418]: DEBUG nova.objects.instance [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'resources' on Instance uuid fe38702e-953d-4748-a523-401e798b2271 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.558914] env[63418]: DEBUG nova.compute.utils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1098.560543] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1098.560742] env[63418]: DEBUG nova.network.neutron [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1098.613099] env[63418]: DEBUG nova.policy [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e9d38547e714839a808ae0c7e7e97d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '381605706a8b4a1e9cca8603c800f1ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1098.619471] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871a782a-1d22-4e07-92d6-4b5f717277a2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.626966] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe93500f-eb3e-4465-a5f0-16586d0555bf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.656138] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af974af4-ee4a-41c7-a6e4-bd5a5e8ef33e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.664244] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b55b39a-1913-420f-9f41-e51c4d3fdb90 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.677293] env[63418]: DEBUG nova.compute.provider_tree [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.890906] env[63418]: DEBUG nova.network.neutron [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Successfully created port: 9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1099.063935] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1099.181266] env[63418]: DEBUG nova.scheduler.client.report [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1099.685674] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.711330] env[63418]: INFO nova.scheduler.client.report [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted allocations for instance fe38702e-953d-4748-a523-401e798b2271 [ 1100.074185] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1100.104824] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.105114] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.105283] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.105494] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.105666] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.105835] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.106126] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.106303] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.106485] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.106684] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.106878] env[63418]: DEBUG nova.virt.hardware [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.107792] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a95e1c-7908-489d-836f-ead1406a5a4a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.118675] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a61cb7-b90f-4fdd-9241-2a620d268ffb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.221030] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c2110857-06d5-44db-8acc-ac7a9e7b8461 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "fe38702e-953d-4748-a523-401e798b2271" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.449s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.313768] env[63418]: DEBUG nova.compute.manager [req-9a85675e-371d-4b82-bcf4-25edea5d6e9e req-ddd7ad8e-42b7-4789-aab9-eec9096188c6 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-vif-plugged-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1100.313768] env[63418]: DEBUG oslo_concurrency.lockutils [req-9a85675e-371d-4b82-bcf4-25edea5d6e9e req-ddd7ad8e-42b7-4789-aab9-eec9096188c6 service nova] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.313768] env[63418]: DEBUG oslo_concurrency.lockutils [req-9a85675e-371d-4b82-bcf4-25edea5d6e9e req-ddd7ad8e-42b7-4789-aab9-eec9096188c6 service nova] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.313768] env[63418]: DEBUG oslo_concurrency.lockutils [req-9a85675e-371d-4b82-bcf4-25edea5d6e9e req-ddd7ad8e-42b7-4789-aab9-eec9096188c6 service nova] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.313768] env[63418]: DEBUG nova.compute.manager [req-9a85675e-371d-4b82-bcf4-25edea5d6e9e req-ddd7ad8e-42b7-4789-aab9-eec9096188c6 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] No waiting events found dispatching network-vif-plugged-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1100.316050] env[63418]: WARNING nova.compute.manager [req-9a85675e-371d-4b82-bcf4-25edea5d6e9e req-ddd7ad8e-42b7-4789-aab9-eec9096188c6 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received unexpected event network-vif-plugged-9d927b3d-91b4-429d-b74d-96a0e1822173 for instance with vm_state building and task_state spawning. [ 1100.472924] env[63418]: DEBUG nova.network.neutron [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Successfully updated port: 9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1100.976371] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.976545] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.976687] env[63418]: DEBUG nova.network.neutron [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1101.061105] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.061382] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.511915] env[63418]: DEBUG nova.network.neutron [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1101.566277] env[63418]: INFO nova.compute.manager [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Detaching volume 45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6 [ 1101.597907] env[63418]: INFO nova.virt.block_device [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Attempting to driver detach volume 45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6 from mountpoint /dev/sdb [ 1101.598181] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1101.598402] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268539', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'name': 'volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98f51fcd-951c-4c16-bc20-efc62ef359f5', 'attached_at': '', 'detached_at': '', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'serial': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1101.599359] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb1c9a3-3f2f-4b15-904c-0e7592017b84 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.626981] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e86f18b-c32a-4432-a638-6ec73cf2c94f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.634475] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcdd97f-f4ad-453e-8dda-14c9ec72407b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.655728] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410dd546-2099-4d43-afab-220965ef8395 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.670597] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] The volume has not been displaced from its original location: [datastore2] volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6/volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1101.675880] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.676812] env[63418]: DEBUG nova.network.neutron [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.678229] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d4c03a4-5c1f-4146-bfb5-07470a107562 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.696761] env[63418]: DEBUG oslo_vmware.api [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1101.696761] env[63418]: value = "task-1245604" [ 1101.696761] env[63418]: _type = "Task" [ 1101.696761] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.704551] env[63418]: DEBUG oslo_vmware.api [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.192411] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.192791] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Instance network_info: |[{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1102.193264] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:91:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d927b3d-91b4-429d-b74d-96a0e1822173', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1102.201507] env[63418]: DEBUG oslo.service.loopingcall [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1102.202305] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1102.206120] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62139f2d-d1f2-4d79-80c3-f8e8d9c15d38 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.226348] env[63418]: DEBUG oslo_vmware.api [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245604, 'name': ReconfigVM_Task, 'duration_secs': 0.229246} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.227672] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1102.232906] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1102.232906] env[63418]: value = "task-1245605" [ 1102.232906] env[63418]: _type = "Task" [ 1102.232906] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.234034] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc71364c-ba76-48c9-8b1a-50edb9d02590 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.251775] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245605, 'name': CreateVM_Task} progress is 10%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.253072] env[63418]: DEBUG oslo_vmware.api [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1102.253072] env[63418]: value = "task-1245606" [ 1102.253072] env[63418]: _type = "Task" [ 1102.253072] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.260819] env[63418]: DEBUG oslo_vmware.api [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245606, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.343552] env[63418]: DEBUG nova.compute.manager [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1102.343762] env[63418]: DEBUG nova.compute.manager [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing instance network info cache due to event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1102.344110] env[63418]: DEBUG oslo_concurrency.lockutils [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.344380] env[63418]: DEBUG oslo_concurrency.lockutils [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.344674] env[63418]: DEBUG nova.network.neutron [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1102.753200] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245605, 'name': CreateVM_Task, 'duration_secs': 0.495836} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.753576] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1102.756660] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.756880] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.757218] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1102.757484] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1aa5eee-896c-41bc-b570-01e7fc69a8b4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.763790] env[63418]: DEBUG oslo_vmware.api [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245606, 'name': ReconfigVM_Task, 'duration_secs': 0.156143} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.765014] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268539', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'name': 'volume-45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98f51fcd-951c-4c16-bc20-efc62ef359f5', 'attached_at': '', 'detached_at': '', 'volume_id': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6', 'serial': '45c2b4f1-375f-46b9-a8af-d6a1cf34a1c6'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1102.767173] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1102.767173] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]521bb768-e0c1-e9df-0acd-cc1988db0bb2" [ 1102.767173] env[63418]: _type = "Task" [ 1102.767173] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.776244] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521bb768-e0c1-e9df-0acd-cc1988db0bb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.044409] env[63418]: DEBUG nova.network.neutron [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updated VIF entry in instance network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1103.044990] env[63418]: DEBUG nova.network.neutron [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.278029] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]521bb768-e0c1-e9df-0acd-cc1988db0bb2, 'name': SearchDatastore_Task, 'duration_secs': 0.013411} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.278293] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.278528] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1103.278772] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.278919] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.279112] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1103.279366] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebf03bf1-c4b5-414f-a45e-ad4124618ed3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.287439] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1103.287439] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1103.288022] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad41132b-b706-4066-ad79-bafbd3793f07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.292685] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1103.292685] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522aba63-1f05-3005-3bcc-80764fa7ff83" [ 1103.292685] env[63418]: _type = "Task" [ 1103.292685] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.299769] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522aba63-1f05-3005-3bcc-80764fa7ff83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.307025] env[63418]: DEBUG nova.objects.instance [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'flavor' on Instance uuid 98f51fcd-951c-4c16-bc20-efc62ef359f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.548015] env[63418]: DEBUG oslo_concurrency.lockutils [req-c0f17def-e759-4fbd-a2fc-67caddf1d1cb req-d08cda46-ea44-4314-b620-392bceda2f46 service nova] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.805071] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]522aba63-1f05-3005-3bcc-80764fa7ff83, 'name': SearchDatastore_Task, 'duration_secs': 0.007902} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.805071] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4900da5-d52e-4c8f-9b15-91ad97f8e17a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.808752] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1103.808752] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525973b6-75d0-84de-1f0e-5bebf755f8ae" [ 1103.808752] env[63418]: _type = "Task" [ 1103.808752] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.818856] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525973b6-75d0-84de-1f0e-5bebf755f8ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.980999] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.315147] env[63418]: DEBUG oslo_concurrency.lockutils [None req-a7a08c90-a981-493b-b318-a320db00d8b1 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.254s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.319496] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525973b6-75d0-84de-1f0e-5bebf755f8ae, 'name': SearchDatastore_Task, 'duration_secs': 0.008788} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.319929] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.320200] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1104.320467] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76dcacfe-dbf5-4c76-baee-d4545b190c15 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.326676] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1104.326676] env[63418]: value = "task-1245607" [ 1104.326676] env[63418]: _type = "Task" [ 1104.326676] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.334501] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.483667] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Getting list of instances from cluster (obj){ [ 1104.483667] env[63418]: value = "domain-c8" [ 1104.483667] env[63418]: _type = "ClusterComputeResource" [ 1104.483667] env[63418]: } {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1104.484799] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6687faa-6cf5-4229-8e4a-b8f77d8e0464 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.497539] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Got total of 3 instances {{(pid=63418) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1104.497731] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Triggering sync for uuid 98f51fcd-951c-4c16-bc20-efc62ef359f5 {{(pid=63418) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10688}} [ 1104.497912] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Triggering sync for uuid 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10688}} [ 1104.498079] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Triggering sync for uuid 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10688}} [ 1104.498447] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.498682] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.499011] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.499213] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.499456] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.500351] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc93e276-b4cd-4e7b-9f0a-646f8ffdde99 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.503374] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2a6ee2-bd3a-4910-9960-d9206909d2e9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.839359] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431899} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.840305] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1104.840434] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1104.840572] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a30a02e-ed86-467b-a949-41c0985d2f07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.847113] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1104.847113] env[63418]: value = "task-1245608" [ 1104.847113] env[63418]: _type = "Task" [ 1104.847113] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.855647] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245608, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.013932] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.515s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.015577] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.334629] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.334907] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.335161] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "98f51fcd-951c-4c16-bc20-efc62ef359f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.335352] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.335523] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.337758] env[63418]: INFO nova.compute.manager [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Terminating instance [ 1105.357490] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245608, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067791} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.357734] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1105.358482] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd9561a-b78a-42a2-927a-e7f4680c02aa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.380103] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.380340] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa9b52c4-f3c5-427c-be0a-5cad142bf1f3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.399481] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1105.399481] env[63418]: value = "task-1245609" [ 1105.399481] env[63418]: _type = "Task" [ 1105.399481] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.406883] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245609, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.841833] env[63418]: DEBUG nova.compute.manager [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1105.842223] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.842978] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55778dbc-8f6c-4cfd-a5f3-051894e3dab4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.850896] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.851147] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b877db49-3307-4e84-8259-0f28a6929378 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.857679] env[63418]: DEBUG oslo_vmware.api [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1105.857679] env[63418]: value = "task-1245610" [ 1105.857679] env[63418]: _type = "Task" [ 1105.857679] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.864905] env[63418]: DEBUG oslo_vmware.api [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.908832] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245609, 'name': ReconfigVM_Task, 'duration_secs': 0.253751} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.909162] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.909818] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4aafb3a7-5a77-4cdd-a6ae-bcf32c1c7578 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.916306] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1105.916306] env[63418]: value = "task-1245611" [ 1105.916306] env[63418]: _type = "Task" [ 1105.916306] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.924429] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245611, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.367132] env[63418]: DEBUG oslo_vmware.api [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245610, 'name': PowerOffVM_Task, 'duration_secs': 0.197619} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.367417] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.367609] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.367866] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e943df61-9384-4d14-9009-4cde329ab96d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.427510] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245611, 'name': Rename_Task, 'duration_secs': 0.153009} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.427778] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1106.428027] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2d8d838-d087-41d3-a70f-cb55d4d6a570 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.433702] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1106.433702] env[63418]: value = "task-1245613" [ 1106.433702] env[63418]: _type = "Task" [ 1106.433702] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.441162] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.442272] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.442476] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.442654] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleting the datastore file [datastore2] 98f51fcd-951c-4c16-bc20-efc62ef359f5 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.442893] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff01acff-fb86-4e5a-a356-81fce62e3651 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.448304] env[63418]: DEBUG oslo_vmware.api [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1106.448304] env[63418]: value = "task-1245614" [ 1106.448304] env[63418]: _type = "Task" [ 1106.448304] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.455929] env[63418]: DEBUG oslo_vmware.api [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.944244] env[63418]: DEBUG oslo_vmware.api [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245613, 'name': PowerOnVM_Task, 'duration_secs': 0.492567} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.944673] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.944754] env[63418]: INFO nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Took 6.87 seconds to spawn the instance on the hypervisor. [ 1106.944883] env[63418]: DEBUG nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1106.945689] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5b6889-40ca-4a94-97a5-33669f1cbad7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.960052] env[63418]: DEBUG oslo_vmware.api [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164679} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.960285] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1106.960465] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1106.960635] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1106.961707] env[63418]: INFO nova.compute.manager [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1106.961707] env[63418]: DEBUG oslo.service.loopingcall [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.961707] env[63418]: DEBUG nova.compute.manager [-] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1106.961707] env[63418]: DEBUG nova.network.neutron [-] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1107.398139] env[63418]: DEBUG nova.compute.manager [req-a4749361-de06-43ce-80d0-530164168cae req-f56a1ac8-2d52-46db-b48e-c649f14149ab service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Received event network-vif-deleted-a6337c68-709c-4b53-b00a-b0e794cd45db {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1107.398474] env[63418]: INFO nova.compute.manager [req-a4749361-de06-43ce-80d0-530164168cae req-f56a1ac8-2d52-46db-b48e-c649f14149ab service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Neutron deleted interface a6337c68-709c-4b53-b00a-b0e794cd45db; detaching it from the instance and deleting it from the info cache [ 1107.398772] env[63418]: DEBUG nova.network.neutron [req-a4749361-de06-43ce-80d0-530164168cae req-f56a1ac8-2d52-46db-b48e-c649f14149ab service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.464350] env[63418]: INFO nova.compute.manager [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Took 13.78 seconds to build instance. [ 1107.875877] env[63418]: DEBUG nova.network.neutron [-] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.902007] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f8eca52-b843-464f-b925-add0eb1cecbe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.914944] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5548ed8f-3136-4311-91f6-f407c665a06c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.945343] env[63418]: DEBUG nova.compute.manager [req-a4749361-de06-43ce-80d0-530164168cae req-f56a1ac8-2d52-46db-b48e-c649f14149ab service nova] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Detach interface failed, port_id=a6337c68-709c-4b53-b00a-b0e794cd45db, reason: Instance 98f51fcd-951c-4c16-bc20-efc62ef359f5 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1107.965681] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9fe3b7f4-1837-4608-987c-5279125efce8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.293s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.965983] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.466s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.966199] env[63418]: INFO nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] During sync_power_state the instance has a pending task (spawning). Skip. [ 1107.966377] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.378525] env[63418]: INFO nova.compute.manager [-] [instance: 98f51fcd-951c-4c16-bc20-efc62ef359f5] Took 1.42 seconds to deallocate network for instance. [ 1108.884959] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.885287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.885522] env[63418]: DEBUG nova.objects.instance [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'resources' on Instance uuid 98f51fcd-951c-4c16-bc20-efc62ef359f5 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.428208] env[63418]: DEBUG nova.compute.manager [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1109.428496] env[63418]: DEBUG nova.compute.manager [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing instance network info cache due to event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1109.428642] env[63418]: DEBUG oslo_concurrency.lockutils [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.428793] env[63418]: DEBUG oslo_concurrency.lockutils [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.428961] env[63418]: DEBUG nova.network.neutron [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1109.448434] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fb86ef-1dd9-4233-b5e6-2361076f1a07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.456070] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04f4251-1af8-4cec-b85d-07e340764f07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.486759] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f26f56-2734-473a-8d7a-13fd33882234 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.494215] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6842f541-0259-42b3-82c2-b598d43ad622 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.507080] env[63418]: DEBUG nova.compute.provider_tree [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.010486] env[63418]: DEBUG nova.scheduler.client.report [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1110.252140] env[63418]: DEBUG nova.network.neutron [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updated VIF entry in instance network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1110.252544] env[63418]: DEBUG nova.network.neutron [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.515903] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.630s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.536467] env[63418]: INFO nova.scheduler.client.report [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted allocations for instance 98f51fcd-951c-4c16-bc20-efc62ef359f5 [ 1110.755561] env[63418]: DEBUG oslo_concurrency.lockutils [req-59cc93a6-0e13-407d-9f5a-bc8038394a31 req-ca624673-8741-4fc0-9bca-86404ca84250 service nova] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.044983] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7622f0d6-84df-4f28-8c34-4f210f9eecd9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "98f51fcd-951c-4c16-bc20-efc62ef359f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.710s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.499519] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.499790] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.002451] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1114.523020] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.523020] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.523421] env[63418]: INFO nova.compute.claims [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1115.575998] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb4bc48-25fa-4be3-94c1-91085a01b558 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.583664] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2194cf1-2560-444d-9da1-d9ad1e7c30bd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.613834] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92be9c1a-0748-4b85-808b-835af91644c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.620693] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0101d6aa-8d95-49f4-ae6c-c1bb4c264bcd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.633213] env[63418]: DEBUG nova.compute.provider_tree [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.136247] env[63418]: DEBUG nova.scheduler.client.report [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1116.641906] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.642312] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1117.147484] env[63418]: DEBUG nova.compute.utils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1117.149113] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1117.149293] env[63418]: DEBUG nova.network.neutron [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1117.201881] env[63418]: DEBUG nova.policy [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b29941866349482fb9e53dcf87cb1845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05ef1c6c74574217817c6ab14a022b91', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1117.467590] env[63418]: DEBUG nova.network.neutron [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Successfully created port: 02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1117.653174] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1118.664136] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1118.689696] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1118.689958] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1118.690137] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1118.690326] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1118.690477] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1118.690629] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1118.690839] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1118.691014] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1118.691189] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1118.691357] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1118.691534] env[63418]: DEBUG nova.virt.hardware [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1118.692419] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4118dfc1-b988-46c9-ae9f-c0d0300b794a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.700069] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2762a6a-2582-49d5-81bd-7f99f5f71a79 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.893715] env[63418]: DEBUG nova.compute.manager [req-22abcb6d-8091-4f73-860b-4e739e49fed8 req-8fd5130f-884a-4610-8e6f-78d844f68135 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Received event network-vif-plugged-02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1118.893961] env[63418]: DEBUG oslo_concurrency.lockutils [req-22abcb6d-8091-4f73-860b-4e739e49fed8 req-8fd5130f-884a-4610-8e6f-78d844f68135 service nova] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.894298] env[63418]: DEBUG oslo_concurrency.lockutils [req-22abcb6d-8091-4f73-860b-4e739e49fed8 req-8fd5130f-884a-4610-8e6f-78d844f68135 service nova] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.894502] env[63418]: DEBUG oslo_concurrency.lockutils [req-22abcb6d-8091-4f73-860b-4e739e49fed8 req-8fd5130f-884a-4610-8e6f-78d844f68135 service nova] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.894683] env[63418]: DEBUG nova.compute.manager [req-22abcb6d-8091-4f73-860b-4e739e49fed8 req-8fd5130f-884a-4610-8e6f-78d844f68135 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] No waiting events found dispatching network-vif-plugged-02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1118.894854] env[63418]: WARNING nova.compute.manager [req-22abcb6d-8091-4f73-860b-4e739e49fed8 req-8fd5130f-884a-4610-8e6f-78d844f68135 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Received unexpected event network-vif-plugged-02a2964c-4337-419a-abc7-a73ef9a4af5f for instance with vm_state building and task_state spawning. [ 1118.952333] env[63418]: DEBUG nova.network.neutron [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Successfully updated port: 02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1119.454916] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.455093] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.455256] env[63418]: DEBUG nova.network.neutron [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1119.729094] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.985866] env[63418]: DEBUG nova.network.neutron [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1120.121016] env[63418]: DEBUG nova.network.neutron [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updating instance_info_cache with network_info: [{"id": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "address": "fa:16:3e:21:b2:ac", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02a2964c-43", "ovs_interfaceid": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.204409] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.624091] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.624301] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Instance network_info: |[{"id": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "address": "fa:16:3e:21:b2:ac", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02a2964c-43", "ovs_interfaceid": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1120.624592] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:b2:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'af454577-0e89-41a3-a9f2-f39716f62fd5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02a2964c-4337-419a-abc7-a73ef9a4af5f', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1120.632793] env[63418]: DEBUG oslo.service.loopingcall [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1120.633018] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1120.633249] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8549b19e-baf4-4f30-872a-4be0f204dc3d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.653949] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1120.653949] env[63418]: value = "task-1245615" [ 1120.653949] env[63418]: _type = "Task" [ 1120.653949] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.661329] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245615, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.923068] env[63418]: DEBUG nova.compute.manager [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Received event network-changed-02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1120.923068] env[63418]: DEBUG nova.compute.manager [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Refreshing instance network info cache due to event network-changed-02a2964c-4337-419a-abc7-a73ef9a4af5f. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1120.923068] env[63418]: DEBUG oslo_concurrency.lockutils [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] Acquiring lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.923450] env[63418]: DEBUG oslo_concurrency.lockutils [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] Acquired lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.923602] env[63418]: DEBUG nova.network.neutron [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Refreshing network info cache for port 02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1121.164305] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245615, 'name': CreateVM_Task, 'duration_secs': 0.321827} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.164494] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.165230] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.165418] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.165758] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1121.166032] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be5bf0cc-304d-4b76-9a0b-6c911d8ad285 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.170597] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1121.170597] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52131aba-2b62-177a-240d-3ab86cb0087c" [ 1121.170597] env[63418]: _type = "Task" [ 1121.170597] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.178522] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52131aba-2b62-177a-240d-3ab86cb0087c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.609475] env[63418]: DEBUG nova.network.neutron [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updated VIF entry in instance network info cache for port 02a2964c-4337-419a-abc7-a73ef9a4af5f. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.609638] env[63418]: DEBUG nova.network.neutron [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updating instance_info_cache with network_info: [{"id": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "address": "fa:16:3e:21:b2:ac", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02a2964c-43", "ovs_interfaceid": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.680059] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52131aba-2b62-177a-240d-3ab86cb0087c, 'name': SearchDatastore_Task, 'duration_secs': 0.013437} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.680358] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.680593] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1121.680831] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.680982] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.681183] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.681438] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a27e1f71-cbbd-444d-a887-7801573351df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.689627] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.689797] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1121.690466] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff440c9d-ddff-475d-87f3-f222697a2bdd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.694842] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1121.694842] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5211f336-a206-e012-d542-c60bfb85711f" [ 1121.694842] env[63418]: _type = "Task" [ 1121.694842] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.701683] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5211f336-a206-e012-d542-c60bfb85711f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.112956] env[63418]: DEBUG oslo_concurrency.lockutils [req-0a9e08ac-8760-47ca-89f8-1154634f5c5d req-1819013d-cb76-4a05-b84b-ff807458c4d6 service nova] Releasing lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.205442] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5211f336-a206-e012-d542-c60bfb85711f, 'name': SearchDatastore_Task, 'duration_secs': 0.008902} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.206218] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a67c43b-68b9-4871-965c-8c70ae139eb2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.209864] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.211237] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1122.211237] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52daffb9-5d18-04b9-4096-0bef51ad4dde" [ 1122.211237] env[63418]: _type = "Task" [ 1122.211237] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.218636] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52daffb9-5d18-04b9-4096-0bef51ad4dde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.722670] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52daffb9-5d18-04b9-4096-0bef51ad4dde, 'name': SearchDatastore_Task, 'duration_secs': 0.008912} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.722940] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.723218] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 38a131e4-22aa-41f1-8e7a-e2ab4660340c/38a131e4-22aa-41f1-8e7a-e2ab4660340c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1122.723473] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-371db53a-018d-4a37-a8d9-ad14af7a9ceb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.729754] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1122.729754] env[63418]: value = "task-1245616" [ 1122.729754] env[63418]: _type = "Task" [ 1122.729754] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.737137] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.210133] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.210501] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 1123.210501] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.238940] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453011} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.239215] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 38a131e4-22aa-41f1-8e7a-e2ab4660340c/38a131e4-22aa-41f1-8e7a-e2ab4660340c.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1123.239428] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1123.239668] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f3a70f1-c77b-4bfe-b05c-7adccba5b290 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.245560] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1123.245560] env[63418]: value = "task-1245617" [ 1123.245560] env[63418]: _type = "Task" [ 1123.245560] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.252417] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.713879] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.714128] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.714307] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.714485] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1123.715396] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca224595-81cc-4948-8014-1492dc30b4f2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.723610] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08888521-ceff-41bf-8d1c-65886cb8f1a9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.737155] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53990852-ad59-4858-9557-a9b99ab30eb5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.743756] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25da9a5-9e9c-4ce1-bfdd-b3b4fc39444f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.753592] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075237} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.777804] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1123.778168] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180215MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1123.778303] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.778491] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.781442] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33404bd-e07d-402c-a831-e3c00827a614 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.803657] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 38a131e4-22aa-41f1-8e7a-e2ab4660340c/38a131e4-22aa-41f1-8e7a-e2ab4660340c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1123.804536] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01f0d4e2-b070-49bb-86d0-9ca1525114f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.824434] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1123.824434] env[63418]: value = "task-1245618" [ 1123.824434] env[63418]: _type = "Task" [ 1123.824434] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.832593] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.333825] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245618, 'name': ReconfigVM_Task, 'duration_secs': 0.309444} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.334196] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 38a131e4-22aa-41f1-8e7a-e2ab4660340c/38a131e4-22aa-41f1-8e7a-e2ab4660340c.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.334666] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3977dbe0-0cae-44ba-8142-561f85396c9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.341013] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1124.341013] env[63418]: value = "task-1245619" [ 1124.341013] env[63418]: _type = "Task" [ 1124.341013] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.348558] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245619, 'name': Rename_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.803926] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 7d532ab0-bf4c-402c-ac81-892930d13d6b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.804128] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 95240ada-686c-484b-9c68-bc9a1f3d3e70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.804265] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 38a131e4-22aa-41f1-8e7a-e2ab4660340c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.804441] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1124.804576] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1124.849935] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245619, 'name': Rename_Task, 'duration_secs': 0.152955} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.850924] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1124.851636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bb99ef-e44f-4926-932d-8ed39a3226d8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.853898] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbbb82c4-94c4-45a4-9755-c6feef154d10 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.859340] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f61d36f-faaf-458f-a57b-8b864beb0575 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.862987] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1124.862987] env[63418]: value = "task-1245620" [ 1124.862987] env[63418]: _type = "Task" [ 1124.862987] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.891500] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d647ebab-6b9a-4982-b2a7-fbf3e0c8d103 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.897149] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245620, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.901479] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197bbe74-ae6c-4bac-a4a7-a823fb1262b8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.914112] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.373226] env[63418]: DEBUG oslo_vmware.api [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245620, 'name': PowerOnVM_Task, 'duration_secs': 0.444889} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.373612] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1125.373740] env[63418]: INFO nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Took 6.71 seconds to spawn the instance on the hypervisor. [ 1125.373893] env[63418]: DEBUG nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1125.374643] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8806c562-27e0-485a-903c-824118bf2e3a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.416877] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1125.891485] env[63418]: INFO nova.compute.manager [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Took 11.38 seconds to build instance. [ 1125.922551] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1125.922747] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.144s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.393326] env[63418]: DEBUG oslo_concurrency.lockutils [None req-c5fcf163-947d-4da7-995d-d1898cdd439d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.893s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.619547] env[63418]: DEBUG nova.compute.manager [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Received event network-changed-02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1126.619798] env[63418]: DEBUG nova.compute.manager [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Refreshing instance network info cache due to event network-changed-02a2964c-4337-419a-abc7-a73ef9a4af5f. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1126.619968] env[63418]: DEBUG oslo_concurrency.lockutils [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] Acquiring lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.620160] env[63418]: DEBUG oslo_concurrency.lockutils [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] Acquired lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.620302] env[63418]: DEBUG nova.network.neutron [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Refreshing network info cache for port 02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.318352] env[63418]: DEBUG nova.network.neutron [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updated VIF entry in instance network info cache for port 02a2964c-4337-419a-abc7-a73ef9a4af5f. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1127.318711] env[63418]: DEBUG nova.network.neutron [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updating instance_info_cache with network_info: [{"id": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "address": "fa:16:3e:21:b2:ac", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02a2964c-43", "ovs_interfaceid": "02a2964c-4337-419a-abc7-a73ef9a4af5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.821748] env[63418]: DEBUG oslo_concurrency.lockutils [req-b54db189-754a-4ee0-832c-0da4f334ec40 req-0c086d01-6337-4f0d-9d86-8b9f0756ae95 service nova] Releasing lock "refresh_cache-38a131e4-22aa-41f1-8e7a-e2ab4660340c" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.923494] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.923653] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 1128.426322] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Didn't find any instances for network info cache update. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10364}} [ 1128.426495] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.426655] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.426812] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.771792] env[63418]: DEBUG oslo_concurrency.lockutils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.772183] env[63418]: DEBUG oslo_concurrency.lockutils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.275426] env[63418]: DEBUG nova.compute.utils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1133.778558] env[63418]: DEBUG oslo_concurrency.lockutils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.841783] env[63418]: DEBUG oslo_concurrency.lockutils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.842079] env[63418]: DEBUG oslo_concurrency.lockutils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.842325] env[63418]: INFO nova.compute.manager [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Attaching volume 88a50bf2-6b53-47d6-8ba0-210e95dd758a to /dev/sdb [ 1134.871782] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d513b6-335c-4140-8c3a-ea536a339f0b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.879334] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ed0ac4-4b99-4507-8318-c596dec47c5d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.892093] env[63418]: DEBUG nova.virt.block_device [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating existing volume attachment record: 8df00499-0f21-49aa-bfd8-d70a4d59b6fc {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1139.455177] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1139.455446] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268552', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'name': 'volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'serial': '88a50bf2-6b53-47d6-8ba0-210e95dd758a'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1139.456347] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17c6a28-c664-49ac-a2a3-9868221f9d5e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.472970] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7d190e-1f24-4bbb-a594-326d9353d8a4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.497093] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a/volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.497407] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d0b0043-d082-4ec7-b150-9a9410535231 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.515967] env[63418]: DEBUG oslo_vmware.api [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1139.515967] env[63418]: value = "task-1245623" [ 1139.515967] env[63418]: _type = "Task" [ 1139.515967] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.523925] env[63418]: DEBUG oslo_vmware.api [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.025853] env[63418]: DEBUG oslo_vmware.api [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245623, 'name': ReconfigVM_Task, 'duration_secs': 0.347094} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.026169] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a/volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1140.030970] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-793baad2-be82-4ed1-83f3-01b9b942f306 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.045808] env[63418]: DEBUG oslo_vmware.api [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1140.045808] env[63418]: value = "task-1245624" [ 1140.045808] env[63418]: _type = "Task" [ 1140.045808] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.054663] env[63418]: DEBUG oslo_vmware.api [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.555266] env[63418]: DEBUG oslo_vmware.api [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245624, 'name': ReconfigVM_Task, 'duration_secs': 0.143435} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.555593] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268552', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'name': 'volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'serial': '88a50bf2-6b53-47d6-8ba0-210e95dd758a'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1141.590121] env[63418]: DEBUG nova.objects.instance [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.095296] env[63418]: DEBUG oslo_concurrency.lockutils [None req-07b63f51-959d-4ce1-a827-bad4b73d00a9 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.253s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.883485] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.883771] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.387470] env[63418]: DEBUG nova.compute.utils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.890712] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.949275] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.949649] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.949799] env[63418]: INFO nova.compute.manager [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Attaching volume 9cd118c7-64e0-4f95-a82e-8d04961d8bb6 to /dev/sdc [ 1144.982346] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cb07a3-be0e-4560-b3a5-a8cdc86705c1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.989252] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.989491] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.989666] env[63418]: INFO nova.compute.manager [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Shelving [ 1144.991424] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0004fcf-7296-433b-a55d-0fd2f13eb649 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.005687] env[63418]: DEBUG nova.virt.block_device [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating existing volume attachment record: bb444d41-b6b3-40a2-98e8-7d2e4aee7411 {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1146.001635] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.001990] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5656aa28-c2a3-473b-8388-0926a269cfe2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.008538] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1146.008538] env[63418]: value = "task-1245626" [ 1146.008538] env[63418]: _type = "Task" [ 1146.008538] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.016457] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245626, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.517565] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245626, 'name': PowerOffVM_Task, 'duration_secs': 0.195773} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.517835] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1146.518644] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1b9dc8-0e31-48ff-ab12-ed441b1df8bb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.536221] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d856a705-3750-4bca-96c1-199e209c78c3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.046082] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Creating Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1147.046426] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7fc0b690-8730-4580-95a6-6c32831db11f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.053777] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1147.053777] env[63418]: value = "task-1245627" [ 1147.053777] env[63418]: _type = "Task" [ 1147.053777] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.061366] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245627, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.563963] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245627, 'name': CreateSnapshot_Task, 'duration_secs': 0.396625} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.564342] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Created Snapshot of the VM instance {{(pid=63418) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1147.565088] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6f1e4e-e707-456a-9934-50b9e2437122 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.081938] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Creating linked-clone VM from snapshot {{(pid=63418) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1148.082299] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-96350e13-594c-45fd-9c2c-c3045f3d6d80 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.090861] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1148.090861] env[63418]: value = "task-1245629" [ 1148.090861] env[63418]: _type = "Task" [ 1148.090861] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.098519] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245629, 'name': CloneVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.602708] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245629, 'name': CloneVM_Task} progress is 94%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.102059] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245629, 'name': CloneVM_Task, 'duration_secs': 0.913825} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.102404] env[63418]: INFO nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Created linked-clone VM from snapshot [ 1149.103057] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f86ff2-a8f1-4ce1-a7b2-cf4adcea03d6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.110471] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Uploading image cebca985-58be-495c-8736-51fa6a23dfda {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1149.131759] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1149.131759] env[63418]: value = "vm-268555" [ 1149.131759] env[63418]: _type = "VirtualMachine" [ 1149.131759] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1149.132025] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cc89cac8-bb3f-4f73-a192-5da89317f261 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.137698] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease: (returnval){ [ 1149.137698] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea546-263e-cf51-25fb-b9595581908f" [ 1149.137698] env[63418]: _type = "HttpNfcLease" [ 1149.137698] env[63418]: } obtained for exporting VM: (result){ [ 1149.137698] env[63418]: value = "vm-268555" [ 1149.137698] env[63418]: _type = "VirtualMachine" [ 1149.137698] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1149.137947] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the lease: (returnval){ [ 1149.137947] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea546-263e-cf51-25fb-b9595581908f" [ 1149.137947] env[63418]: _type = "HttpNfcLease" [ 1149.137947] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1149.143435] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1149.143435] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea546-263e-cf51-25fb-b9595581908f" [ 1149.143435] env[63418]: _type = "HttpNfcLease" [ 1149.143435] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1149.548488] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1149.548791] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268553', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'name': 'volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'serial': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1149.549760] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8ab6c7-cccc-4097-8e13-4127a053e344 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.565927] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1eeeefd-0225-4f2a-9e81-f5856a6187d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.592362] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6/volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.592595] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc5dcb25-7840-4d22-9de4-b49117a518a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.611210] env[63418]: DEBUG oslo_vmware.api [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1149.611210] env[63418]: value = "task-1245631" [ 1149.611210] env[63418]: _type = "Task" [ 1149.611210] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.620029] env[63418]: DEBUG oslo_vmware.api [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245631, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.645325] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1149.645325] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea546-263e-cf51-25fb-b9595581908f" [ 1149.645325] env[63418]: _type = "HttpNfcLease" [ 1149.645325] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1149.645713] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1149.645713] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]522ea546-263e-cf51-25fb-b9595581908f" [ 1149.645713] env[63418]: _type = "HttpNfcLease" [ 1149.645713] env[63418]: }. {{(pid=63418) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1149.646468] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2331a260-a1e2-4a1b-9999-dccaef8abd87 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.654065] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e4f286-3996-a8c6-a66b-8df72a59ca6e/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1149.654301] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e4f286-3996-a8c6-a66b-8df72a59ca6e/disk-0.vmdk for reading. {{(pid=63418) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1149.743638] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-42993474-8ae0-4109-b1c2-76ab6fee69cb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.122095] env[63418]: DEBUG oslo_vmware.api [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245631, 'name': ReconfigVM_Task, 'duration_secs': 0.364858} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.123032] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6/volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.128093] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ee58f13-fa0a-4055-a030-c435770371df {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.143038] env[63418]: DEBUG oslo_vmware.api [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1150.143038] env[63418]: value = "task-1245632" [ 1150.143038] env[63418]: _type = "Task" [ 1150.143038] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.151797] env[63418]: DEBUG oslo_vmware.api [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.653591] env[63418]: DEBUG oslo_vmware.api [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245632, 'name': ReconfigVM_Task, 'duration_secs': 0.143432} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.654044] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268553', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'name': 'volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'serial': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1151.690048] env[63418]: DEBUG nova.objects.instance [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.195351] env[63418]: DEBUG oslo_concurrency.lockutils [None req-3de674ad-0d81-4199-8e47-925dc157d139 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.492559] env[63418]: DEBUG oslo_concurrency.lockutils [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.492559] env[63418]: DEBUG oslo_concurrency.lockutils [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.996055] env[63418]: INFO nova.compute.manager [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Detaching volume 88a50bf2-6b53-47d6-8ba0-210e95dd758a [ 1153.027098] env[63418]: INFO nova.virt.block_device [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Attempting to driver detach volume 88a50bf2-6b53-47d6-8ba0-210e95dd758a from mountpoint /dev/sdb [ 1153.027396] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1153.027651] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268552', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'name': 'volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'serial': '88a50bf2-6b53-47d6-8ba0-210e95dd758a'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1153.028605] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05fe34e-43ae-4a6e-8761-87e9759ed710 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.056980] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c12787-e3ab-4950-9461-0278796a764c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.064873] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6965ae-4a78-4fee-ac25-d8d206d971fb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.090503] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7581b47f-3e13-489a-b599-370add13a79f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.107677] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] The volume has not been displaced from its original location: [datastore2] volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a/volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1153.112970] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1153.113312] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe62061d-85ef-4fe3-8f64-8fe2388e6cdf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.131656] env[63418]: DEBUG oslo_vmware.api [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1153.131656] env[63418]: value = "task-1245633" [ 1153.131656] env[63418]: _type = "Task" [ 1153.131656] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.140047] env[63418]: DEBUG oslo_vmware.api [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.642620] env[63418]: DEBUG oslo_vmware.api [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245633, 'name': ReconfigVM_Task, 'duration_secs': 0.290683} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.642911] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1153.647561] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-038f9a21-49e9-4be6-9fff-707e17468fb3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.662219] env[63418]: DEBUG oslo_vmware.api [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1153.662219] env[63418]: value = "task-1245634" [ 1153.662219] env[63418]: _type = "Task" [ 1153.662219] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.672112] env[63418]: DEBUG oslo_vmware.api [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245634, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.172915] env[63418]: DEBUG oslo_vmware.api [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245634, 'name': ReconfigVM_Task, 'duration_secs': 0.233142} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.173337] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268552', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'name': 'volume-88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '88a50bf2-6b53-47d6-8ba0-210e95dd758a', 'serial': '88a50bf2-6b53-47d6-8ba0-210e95dd758a'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1154.713814] env[63418]: DEBUG nova.objects.instance [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.723061] env[63418]: DEBUG oslo_concurrency.lockutils [None req-533f59ae-3cf1-4a5f-9c6f-5fd6402fe64b tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.231s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.732705] env[63418]: DEBUG oslo_concurrency.lockutils [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.732991] env[63418]: DEBUG oslo_concurrency.lockutils [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.236742] env[63418]: INFO nova.compute.manager [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Detaching volume 9cd118c7-64e0-4f95-a82e-8d04961d8bb6 [ 1156.270055] env[63418]: INFO nova.virt.block_device [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Attempting to driver detach volume 9cd118c7-64e0-4f95-a82e-8d04961d8bb6 from mountpoint /dev/sdc [ 1156.270328] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1156.270523] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268553', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'name': 'volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'serial': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1156.271427] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a98ab76-c9aa-4cf8-800e-77ccc1b90ebc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.295223] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb88e76-0def-4be5-8d75-d93994e5a53e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.302770] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cbcd42-ff08-48e1-99be-e89bd20a5f29 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.325922] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5adf247-3e61-4e3e-a729-1c8ee39a3bf4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.342297] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] The volume has not been displaced from its original location: [datastore1] volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6/volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1156.347798] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfiguring VM instance instance-0000006e to detach disk 2002 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1156.348173] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ce4a5ff-22e4-4e06-9fc7-637daa6946a8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.366682] env[63418]: DEBUG oslo_vmware.api [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1156.366682] env[63418]: value = "task-1245635" [ 1156.366682] env[63418]: _type = "Task" [ 1156.366682] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.374916] env[63418]: DEBUG oslo_vmware.api [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245635, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.877234] env[63418]: DEBUG oslo_vmware.api [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245635, 'name': ReconfigVM_Task, 'duration_secs': 0.273454} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.877609] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Reconfigured VM instance instance-0000006e to detach disk 2002 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1156.882542] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45e22e3a-f7f9-4663-a0bf-53dbaf45f489 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.898026] env[63418]: DEBUG oslo_vmware.api [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1156.898026] env[63418]: value = "task-1245636" [ 1156.898026] env[63418]: _type = "Task" [ 1156.898026] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.906640] env[63418]: DEBUG oslo_vmware.api [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245636, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.408497] env[63418]: DEBUG oslo_vmware.api [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245636, 'name': ReconfigVM_Task, 'duration_secs': 0.149993} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.408856] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268553', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'name': 'volume-9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d532ab0-bf4c-402c-ac81-892930d13d6b', 'attached_at': '', 'detached_at': '', 'volume_id': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6', 'serial': '9cd118c7-64e0-4f95-a82e-8d04961d8bb6'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1157.848781] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e4f286-3996-a8c6-a66b-8df72a59ca6e/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1157.849724] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a66c15-33e1-4652-a847-67d5e0ef4159 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.855481] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e4f286-3996-a8c6-a66b-8df72a59ca6e/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1157.855675] env[63418]: ERROR oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e4f286-3996-a8c6-a66b-8df72a59ca6e/disk-0.vmdk due to incomplete transfer. [ 1157.855897] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f486827e-4f20-4214-8028-a253bcbcb49a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.861837] env[63418]: DEBUG oslo_vmware.rw_handles [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e4f286-3996-a8c6-a66b-8df72a59ca6e/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1157.862057] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Uploaded image cebca985-58be-495c-8736-51fa6a23dfda to the Glance image server {{(pid=63418) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1157.864268] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Destroying the VM {{(pid=63418) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1157.864492] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4e7a1371-f207-4ef6-85af-63caafdb65cc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.869214] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1157.869214] env[63418]: value = "task-1245637" [ 1157.869214] env[63418]: _type = "Task" [ 1157.869214] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.876526] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245637, 'name': Destroy_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.953535] env[63418]: DEBUG nova.objects.instance [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'flavor' on Instance uuid 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.378824] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245637, 'name': Destroy_Task, 'duration_secs': 0.308811} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.379140] env[63418]: INFO nova.virt.vmwareapi.vm_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Destroyed the VM [ 1158.379456] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deleting Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1158.379716] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-99898bdd-324a-4401-982f-64168ca3fb8b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.385166] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1158.385166] env[63418]: value = "task-1245638" [ 1158.385166] env[63418]: _type = "Task" [ 1158.385166] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.392493] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245638, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.894652] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245638, 'name': RemoveSnapshot_Task, 'duration_secs': 0.328538} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.894934] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deleted Snapshot of the VM instance {{(pid=63418) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1158.895203] env[63418]: DEBUG nova.compute.manager [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1158.895925] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c842ba5c-2698-4ae5-8c55-2b0094181173 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.960926] env[63418]: DEBUG oslo_concurrency.lockutils [None req-331686a1-a0d1-41dd-bd61-cbcec2139cc8 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.228s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.407082] env[63418]: INFO nova.compute.manager [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Shelve offloading [ 1159.910656] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1159.911038] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-862ba168-aa17-4e13-ab35-cf4653ddf2f9 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.919016] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1159.919016] env[63418]: value = "task-1245639" [ 1159.919016] env[63418]: _type = "Task" [ 1159.919016] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.926555] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.110559] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.110948] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.111173] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "7d532ab0-bf4c-402c-ac81-892930d13d6b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.111411] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.111633] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.113861] env[63418]: INFO nova.compute.manager [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Terminating instance [ 1160.429591] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] VM already powered off {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1160.429813] env[63418]: DEBUG nova.compute.manager [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1160.430573] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc861702-39f4-4623-bd4b-ad00af2f9e92 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.435794] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.435962] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.436149] env[63418]: DEBUG nova.network.neutron [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1160.618177] env[63418]: DEBUG nova.compute.manager [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1160.618407] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1160.619317] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2660d119-1586-4bee-878c-0ec4698729a0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.626527] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1160.626768] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70987b2e-27d5-4089-9e49-d767c51f5e54 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.632766] env[63418]: DEBUG oslo_vmware.api [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1160.632766] env[63418]: value = "task-1245640" [ 1160.632766] env[63418]: _type = "Task" [ 1160.632766] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.641820] env[63418]: DEBUG oslo_vmware.api [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245640, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.137323] env[63418]: DEBUG nova.network.neutron [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.143588] env[63418]: DEBUG oslo_vmware.api [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245640, 'name': PowerOffVM_Task, 'duration_secs': 0.148774} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.143844] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1161.144035] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1161.144289] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43be8dd9-d983-4ec5-b276-cec2666b63a7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.214244] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1161.214475] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1161.214661] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Deleting the datastore file [datastore2] 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1161.214938] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7c04734-bd12-4674-85f4-ad5fe11eb343 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.221522] env[63418]: DEBUG oslo_vmware.api [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for the task: (returnval){ [ 1161.221522] env[63418]: value = "task-1245642" [ 1161.221522] env[63418]: _type = "Task" [ 1161.221522] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.228527] env[63418]: DEBUG oslo_vmware.api [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245642, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.640617] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.731154] env[63418]: DEBUG oslo_vmware.api [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Task: {'id': task-1245642, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121065} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.731428] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1161.731611] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1161.731790] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1161.731972] env[63418]: INFO nova.compute.manager [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1161.732226] env[63418]: DEBUG oslo.service.loopingcall [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.732425] env[63418]: DEBUG nova.compute.manager [-] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1161.732517] env[63418]: DEBUG nova.network.neutron [-] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1161.885578] env[63418]: DEBUG nova.compute.manager [req-99d8e6b3-6c24-42af-85d1-e1a467fad7be req-5fddcab4-0a29-4bca-917d-f9efa1a1bd25 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-vif-unplugged-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1161.885954] env[63418]: DEBUG oslo_concurrency.lockutils [req-99d8e6b3-6c24-42af-85d1-e1a467fad7be req-5fddcab4-0a29-4bca-917d-f9efa1a1bd25 service nova] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.886727] env[63418]: DEBUG oslo_concurrency.lockutils [req-99d8e6b3-6c24-42af-85d1-e1a467fad7be req-5fddcab4-0a29-4bca-917d-f9efa1a1bd25 service nova] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.887034] env[63418]: DEBUG oslo_concurrency.lockutils [req-99d8e6b3-6c24-42af-85d1-e1a467fad7be req-5fddcab4-0a29-4bca-917d-f9efa1a1bd25 service nova] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.887327] env[63418]: DEBUG nova.compute.manager [req-99d8e6b3-6c24-42af-85d1-e1a467fad7be req-5fddcab4-0a29-4bca-917d-f9efa1a1bd25 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] No waiting events found dispatching network-vif-unplugged-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1161.887661] env[63418]: WARNING nova.compute.manager [req-99d8e6b3-6c24-42af-85d1-e1a467fad7be req-5fddcab4-0a29-4bca-917d-f9efa1a1bd25 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received unexpected event network-vif-unplugged-9d927b3d-91b4-429d-b74d-96a0e1822173 for instance with vm_state shelved and task_state shelving_offloading. [ 1161.987180] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1161.988104] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7d49f0-17c1-4883-b8fa-bca878c1a4d2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.997914] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1161.998170] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7fad459-22c6-414a-b679-8b697096636b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.082330] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1162.082492] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1162.082736] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleting the datastore file [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.082960] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af1de21f-7e50-48e8-8c3c-15e37f41805b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.091114] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1162.091114] env[63418]: value = "task-1245644" [ 1162.091114] env[63418]: _type = "Task" [ 1162.091114] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.099194] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245644, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.207224] env[63418]: DEBUG nova.compute.manager [req-dc335aac-a2e6-4acb-85d3-54aa9ad94fee req-4c00e70a-7a36-4771-b30a-ceed39c5cf32 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Received event network-vif-deleted-22836017-f955-4526-9615-0c23c172ddf2 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1162.207499] env[63418]: INFO nova.compute.manager [req-dc335aac-a2e6-4acb-85d3-54aa9ad94fee req-4c00e70a-7a36-4771-b30a-ceed39c5cf32 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Neutron deleted interface 22836017-f955-4526-9615-0c23c172ddf2; detaching it from the instance and deleting it from the info cache [ 1162.207677] env[63418]: DEBUG nova.network.neutron [req-dc335aac-a2e6-4acb-85d3-54aa9ad94fee req-4c00e70a-7a36-4771-b30a-ceed39c5cf32 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.601063] env[63418]: DEBUG oslo_vmware.api [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245644, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129705} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.601358] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1162.601583] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1162.601743] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1162.623356] env[63418]: INFO nova.scheduler.client.report [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted allocations for instance 95240ada-686c-484b-9c68-bc9a1f3d3e70 [ 1162.686398] env[63418]: DEBUG nova.network.neutron [-] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.710089] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3290a64e-f7a8-4df9-b4ba-2f6bd1452e89 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.719808] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7da319-241e-4658-a930-6cbe8e7dbfc0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.743113] env[63418]: DEBUG nova.compute.manager [req-dc335aac-a2e6-4acb-85d3-54aa9ad94fee req-4c00e70a-7a36-4771-b30a-ceed39c5cf32 service nova] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Detach interface failed, port_id=22836017-f955-4526-9615-0c23c172ddf2, reason: Instance 7d532ab0-bf4c-402c-ac81-892930d13d6b could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1163.128854] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.129287] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.129649] env[63418]: DEBUG nova.objects.instance [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'resources' on Instance uuid 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.189431] env[63418]: INFO nova.compute.manager [-] [instance: 7d532ab0-bf4c-402c-ac81-892930d13d6b] Took 1.46 seconds to deallocate network for instance. [ 1163.345080] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.345343] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.632513] env[63418]: DEBUG nova.objects.instance [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'numa_topology' on Instance uuid 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.696504] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.848222] env[63418]: DEBUG nova.compute.utils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1163.954379] env[63418]: DEBUG nova.compute.manager [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1163.954501] env[63418]: DEBUG nova.compute.manager [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing instance network info cache due to event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1163.954714] env[63418]: DEBUG oslo_concurrency.lockutils [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.954869] env[63418]: DEBUG oslo_concurrency.lockutils [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.955044] env[63418]: DEBUG nova.network.neutron [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1164.134907] env[63418]: DEBUG nova.objects.base [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Object Instance<95240ada-686c-484b-9c68-bc9a1f3d3e70> lazy-loaded attributes: resources,numa_topology {{(pid=63418) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1164.168613] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf9a742-d12f-43a2-8b97-c7ff04635cb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.175773] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992ef2a8-4681-475b-9619-08ce2d432097 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.204167] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a58aa4-b6a3-450a-97fd-4c3cf151a224 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.211037] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66e63c1-8018-49b2-8831-276d339e539e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.224430] env[63418]: DEBUG nova.compute.provider_tree [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.351496] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.670497] env[63418]: DEBUG nova.network.neutron [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updated VIF entry in instance network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1164.670856] env[63418]: DEBUG nova.network.neutron [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": null, "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9d927b3d-91", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.727463] env[63418]: DEBUG nova.scheduler.client.report [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1165.088269] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.173910] env[63418]: DEBUG oslo_concurrency.lockutils [req-65333d03-68cf-4958-b117-d5cd1b1524dd req-371744a5-8ebf-4876-b1b9-2bb1dfa5ff6c service nova] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.232626] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.234968] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.539s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.235232] env[63418]: DEBUG nova.objects.instance [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lazy-loading 'resources' on Instance uuid 7d532ab0-bf4c-402c-ac81-892930d13d6b {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.412178] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.412540] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.412635] env[63418]: INFO nova.compute.manager [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Attaching volume aa6aa914-922b-4a18-b707-86a15e15d18b to /dev/sdb [ 1165.441998] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e879c84f-7f6c-4114-b650-9eb2c61baf97 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.449235] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8500b4ec-8dda-420a-8f8f-b35cfdf14255 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.461606] env[63418]: DEBUG nova.virt.block_device [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updating existing volume attachment record: d13a2c02-c530-4f2b-9b33-9ea7117ef558 {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1165.743705] env[63418]: DEBUG oslo_concurrency.lockutils [None req-1b2da993-e19f-465e-aae3-61808f918cd8 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.754s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.744576] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.657s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.744762] env[63418]: INFO nova.compute.manager [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Unshelving [ 1165.791748] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e12311-7643-4a48-b090-45fb506d5bae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.799071] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11e0fcb-6dae-4dbe-9f5b-404cb0c56f92 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.829543] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3fad7a-5505-4aae-802e-3416a8a426b1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.836051] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2850bd39-1a90-4cb8-bdd0-f17806a0b192 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.848577] env[63418]: DEBUG nova.compute.provider_tree [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.351317] env[63418]: DEBUG nova.scheduler.client.report [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1166.768987] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.855631] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.858040] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.089s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.858161] env[63418]: DEBUG nova.objects.instance [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'pci_requests' on Instance uuid 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.873205] env[63418]: INFO nova.scheduler.client.report [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Deleted allocations for instance 7d532ab0-bf4c-402c-ac81-892930d13d6b [ 1167.362222] env[63418]: DEBUG nova.objects.instance [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'numa_topology' on Instance uuid 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.380811] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7939d225-0fc2-4894-83ac-1940b9c23262 tempest-AttachVolumeTestJSON-1671337528 tempest-AttachVolumeTestJSON-1671337528-project-member] Lock "7d532ab0-bf4c-402c-ac81-892930d13d6b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.270s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.865333] env[63418]: INFO nova.compute.claims [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1168.913686] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e12ecda-c7d8-44bf-845d-3696d1d52248 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.920636] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a4673f-8da6-4bc1-8906-043f9578b77f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.950726] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07dc2a06-67f1-4098-aa28-8f0028393b6d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.957357] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0aa6653-1bbd-4eee-adad-394a285fea9a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.970269] env[63418]: DEBUG nova.compute.provider_tree [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.473296] env[63418]: DEBUG nova.scheduler.client.report [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1169.978870] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.121s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.005121] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1170.005433] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268556', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'name': 'volume-aa6aa914-922b-4a18-b707-86a15e15d18b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38a131e4-22aa-41f1-8e7a-e2ab4660340c', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'serial': 'aa6aa914-922b-4a18-b707-86a15e15d18b'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1170.007225] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6346bdca-c519-4e48-83d2-76b93b0e2fba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.010728] env[63418]: INFO nova.network.neutron [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating port 9d927b3d-91b4-429d-b74d-96a0e1822173 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1170.027149] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c046d7-46cd-428d-8218-e2f85880a3c4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.053741] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] volume-aa6aa914-922b-4a18-b707-86a15e15d18b/volume-aa6aa914-922b-4a18-b707-86a15e15d18b.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1170.054048] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be1f6725-8f9e-40b1-acea-3e0e0cad4150 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.073062] env[63418]: DEBUG oslo_vmware.api [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1170.073062] env[63418]: value = "task-1245648" [ 1170.073062] env[63418]: _type = "Task" [ 1170.073062] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.081519] env[63418]: DEBUG oslo_vmware.api [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.583382] env[63418]: DEBUG oslo_vmware.api [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245648, 'name': ReconfigVM_Task, 'duration_secs': 0.351475} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.583691] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Reconfigured VM instance instance-00000070 to attach disk [datastore2] volume-aa6aa914-922b-4a18-b707-86a15e15d18b/volume-aa6aa914-922b-4a18-b707-86a15e15d18b.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1170.588453] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e30679a-8284-40a2-af5a-bb2774beac73 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.603581] env[63418]: DEBUG oslo_vmware.api [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1170.603581] env[63418]: value = "task-1245649" [ 1170.603581] env[63418]: _type = "Task" [ 1170.603581] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.612811] env[63418]: DEBUG oslo_vmware.api [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245649, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.113738] env[63418]: DEBUG oslo_vmware.api [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245649, 'name': ReconfigVM_Task, 'duration_secs': 0.145321} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.115023] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268556', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'name': 'volume-aa6aa914-922b-4a18-b707-86a15e15d18b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38a131e4-22aa-41f1-8e7a-e2ab4660340c', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'serial': 'aa6aa914-922b-4a18-b707-86a15e15d18b'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1171.398037] env[63418]: DEBUG nova.compute.manager [req-8ac5ff53-7f3a-441b-988d-bcb4f9a753ae req-9a519dc2-e54a-4af1-bd2c-bc49410921c7 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-vif-plugged-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1171.398270] env[63418]: DEBUG oslo_concurrency.lockutils [req-8ac5ff53-7f3a-441b-988d-bcb4f9a753ae req-9a519dc2-e54a-4af1-bd2c-bc49410921c7 service nova] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.398486] env[63418]: DEBUG oslo_concurrency.lockutils [req-8ac5ff53-7f3a-441b-988d-bcb4f9a753ae req-9a519dc2-e54a-4af1-bd2c-bc49410921c7 service nova] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.398655] env[63418]: DEBUG oslo_concurrency.lockutils [req-8ac5ff53-7f3a-441b-988d-bcb4f9a753ae req-9a519dc2-e54a-4af1-bd2c-bc49410921c7 service nova] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.398874] env[63418]: DEBUG nova.compute.manager [req-8ac5ff53-7f3a-441b-988d-bcb4f9a753ae req-9a519dc2-e54a-4af1-bd2c-bc49410921c7 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] No waiting events found dispatching network-vif-plugged-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1171.399076] env[63418]: WARNING nova.compute.manager [req-8ac5ff53-7f3a-441b-988d-bcb4f9a753ae req-9a519dc2-e54a-4af1-bd2c-bc49410921c7 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received unexpected event network-vif-plugged-9d927b3d-91b4-429d-b74d-96a0e1822173 for instance with vm_state shelved_offloaded and task_state spawning. [ 1171.480309] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.480506] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.480693] env[63418]: DEBUG nova.network.neutron [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1172.150014] env[63418]: DEBUG nova.objects.instance [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'flavor' on Instance uuid 38a131e4-22aa-41f1-8e7a-e2ab4660340c {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.174728] env[63418]: DEBUG nova.network.neutron [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.654645] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7ced65ee-b263-4077-8787-e1cd2e7e72ff tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.242s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.677105] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.701579] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8a3045c3f61e85c91db4b7887b7807ec',container_format='bare',created_at=2024-10-10T13:48:27Z,direct_url=,disk_format='vmdk',id=cebca985-58be-495c-8736-51fa6a23dfda,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-2006035673-shelved',owner='381605706a8b4a1e9cca8603c800f1ac',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-10T13:48:40Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1172.701871] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1172.702096] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1172.702338] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1172.702531] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1172.702792] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1172.703085] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1172.703302] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1172.703518] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1172.703730] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1172.703954] env[63418]: DEBUG nova.virt.hardware [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1172.705166] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9023384-a294-4242-ae63-25d99e647ab2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.713902] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8425593-54b2-472b-9ad2-d9603c569a07 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.728187] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:91:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d927b3d-91b4-429d-b74d-96a0e1822173', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1172.736421] env[63418]: DEBUG oslo.service.loopingcall [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1172.736690] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1172.737205] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3edfbeb9-ce46-4b77-bbc7-cfce5a97c84a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.757756] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1172.757756] env[63418]: value = "task-1245651" [ 1172.757756] env[63418]: _type = "Task" [ 1172.757756] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.765550] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245651, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.853662] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.853989] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.267487] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245651, 'name': CreateVM_Task, 'duration_secs': 0.305888} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.267848] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1173.268309] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.268489] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.268917] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1173.269200] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58b027f7-a0e5-4638-96dd-806b0a3df21d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.273396] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1173.273396] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5270bcb4-0c98-2d53-e4d5-bab4a74908d3" [ 1173.273396] env[63418]: _type = "Task" [ 1173.273396] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.280702] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5270bcb4-0c98-2d53-e4d5-bab4a74908d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.357595] env[63418]: INFO nova.compute.manager [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Detaching volume aa6aa914-922b-4a18-b707-86a15e15d18b [ 1173.388519] env[63418]: INFO nova.virt.block_device [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Attempting to driver detach volume aa6aa914-922b-4a18-b707-86a15e15d18b from mountpoint /dev/sdb [ 1173.388750] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1173.388990] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268556', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'name': 'volume-aa6aa914-922b-4a18-b707-86a15e15d18b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38a131e4-22aa-41f1-8e7a-e2ab4660340c', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'serial': 'aa6aa914-922b-4a18-b707-86a15e15d18b'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1173.389849] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f409b8-589b-4588-a98e-947d638e3cae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.410585] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a90e28-192c-4364-bde2-172a0f4ae5bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.417066] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3405a035-bf1c-48d2-a7dd-8601b09c205e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.422173] env[63418]: DEBUG nova.compute.manager [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1173.422367] env[63418]: DEBUG nova.compute.manager [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing instance network info cache due to event network-changed-9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1173.422590] env[63418]: DEBUG oslo_concurrency.lockutils [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] Acquiring lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.422737] env[63418]: DEBUG oslo_concurrency.lockutils [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] Acquired lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.422900] env[63418]: DEBUG nova.network.neutron [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Refreshing network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1173.441074] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8ce6f3-7464-4142-9c84-013ee808e921 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.456106] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] The volume has not been displaced from its original location: [datastore2] volume-aa6aa914-922b-4a18-b707-86a15e15d18b/volume-aa6aa914-922b-4a18-b707-86a15e15d18b.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1173.461264] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1173.461724] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7179369-63c4-44b1-919d-075dd7fe938b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.479243] env[63418]: DEBUG oslo_vmware.api [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1173.479243] env[63418]: value = "task-1245652" [ 1173.479243] env[63418]: _type = "Task" [ 1173.479243] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.487174] env[63418]: DEBUG oslo_vmware.api [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245652, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.783852] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.784149] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Processing image cebca985-58be-495c-8736-51fa6a23dfda {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1173.784451] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.784609] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.784795] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1173.785094] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0283e59f-fd8a-481c-b6a7-bb46af4df2ae {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.794681] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1173.794860] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1173.795568] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f3ba148-da47-43b9-b402-5c23ff97023f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.800217] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1173.800217] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52156e7e-0374-4111-2c8a-6745f2e89111" [ 1173.800217] env[63418]: _type = "Task" [ 1173.800217] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.806971] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]52156e7e-0374-4111-2c8a-6745f2e89111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.989114] env[63418]: DEBUG oslo_vmware.api [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245652, 'name': ReconfigVM_Task, 'duration_secs': 0.21387} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.989410] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1173.996202] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e70466f0-995a-4bbc-bd7c-cb120764d9c6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.010529] env[63418]: DEBUG oslo_vmware.api [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1174.010529] env[63418]: value = "task-1245654" [ 1174.010529] env[63418]: _type = "Task" [ 1174.010529] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.018500] env[63418]: DEBUG oslo_vmware.api [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245654, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.173956] env[63418]: DEBUG nova.network.neutron [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updated VIF entry in instance network info cache for port 9d927b3d-91b4-429d-b74d-96a0e1822173. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1174.174379] env[63418]: DEBUG nova.network.neutron [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [{"id": "9d927b3d-91b4-429d-b74d-96a0e1822173", "address": "fa:16:3e:45:91:0b", "network": {"id": "e4d140da-5dd8-4e1f-a15b-54e44dbbdc6c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-109778299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "381605706a8b4a1e9cca8603c800f1ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d927b3d-91", "ovs_interfaceid": "9d927b3d-91b4-429d-b74d-96a0e1822173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.310760] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Preparing fetch location {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1174.311125] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Fetch image to [datastore2] OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260/OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260.vmdk {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1174.311252] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Downloading stream optimized image cebca985-58be-495c-8736-51fa6a23dfda to [datastore2] OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260/OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260.vmdk on the data store datastore2 as vApp {{(pid=63418) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1174.311463] env[63418]: DEBUG nova.virt.vmwareapi.images [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Downloading image file data cebca985-58be-495c-8736-51fa6a23dfda to the ESX as VM named 'OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260' {{(pid=63418) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1174.379543] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1174.379543] env[63418]: value = "resgroup-9" [ 1174.379543] env[63418]: _type = "ResourcePool" [ 1174.379543] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1174.379828] env[63418]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a4d91587-2791-44d3-a53b-75a76542aeaa {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.399357] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease: (returnval){ [ 1174.399357] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abdb9d-a889-94d0-3ca6-5403f2ff99d7" [ 1174.399357] env[63418]: _type = "HttpNfcLease" [ 1174.399357] env[63418]: } obtained for vApp import into resource pool (val){ [ 1174.399357] env[63418]: value = "resgroup-9" [ 1174.399357] env[63418]: _type = "ResourcePool" [ 1174.399357] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1174.399640] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the lease: (returnval){ [ 1174.399640] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abdb9d-a889-94d0-3ca6-5403f2ff99d7" [ 1174.399640] env[63418]: _type = "HttpNfcLease" [ 1174.399640] env[63418]: } to be ready. {{(pid=63418) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1174.405299] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1174.405299] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abdb9d-a889-94d0-3ca6-5403f2ff99d7" [ 1174.405299] env[63418]: _type = "HttpNfcLease" [ 1174.405299] env[63418]: } is initializing. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1174.520614] env[63418]: DEBUG oslo_vmware.api [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245654, 'name': ReconfigVM_Task, 'duration_secs': 0.127165} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.520956] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268556', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'name': 'volume-aa6aa914-922b-4a18-b707-86a15e15d18b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38a131e4-22aa-41f1-8e7a-e2ab4660340c', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa6aa914-922b-4a18-b707-86a15e15d18b', 'serial': 'aa6aa914-922b-4a18-b707-86a15e15d18b'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1174.677491] env[63418]: DEBUG oslo_concurrency.lockutils [req-1ffae5a5-7582-4cb2-a841-258c6267b2ab req-bc9614ae-fe32-4185-9746-7c37e3374576 service nova] Releasing lock "refresh_cache-95240ada-686c-484b-9c68-bc9a1f3d3e70" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.908247] env[63418]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1174.908247] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abdb9d-a889-94d0-3ca6-5403f2ff99d7" [ 1174.908247] env[63418]: _type = "HttpNfcLease" [ 1174.908247] env[63418]: } is ready. {{(pid=63418) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1174.908550] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1174.908550] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]52abdb9d-a889-94d0-3ca6-5403f2ff99d7" [ 1174.908550] env[63418]: _type = "HttpNfcLease" [ 1174.908550] env[63418]: }. {{(pid=63418) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1174.909326] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e973f3-dd60-4beb-b15c-9bcd779cee74 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.916720] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c0862-fd71-cd60-92bd-04815450d0c4/disk-0.vmdk from lease info. {{(pid=63418) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1174.916873] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c0862-fd71-cd60-92bd-04815450d0c4/disk-0.vmdk. {{(pid=63418) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1174.981129] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cd8379a7-2dd9-497e-bbcc-d1709a2ccaf5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.065598] env[63418]: DEBUG nova.objects.instance [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'flavor' on Instance uuid 38a131e4-22aa-41f1-8e7a-e2ab4660340c {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.066401] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Completed reading data from the image iterator. {{(pid=63418) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1176.066657] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c0862-fd71-cd60-92bd-04815450d0c4/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1176.067602] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a50fdc-657a-4eea-ae09-2283427291e8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.074135] env[63418]: DEBUG oslo_concurrency.lockutils [None req-e04842fb-543f-4afe-a64d-f294fbcd8e9a tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.220s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.075166] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c0862-fd71-cd60-92bd-04815450d0c4/disk-0.vmdk is in state: ready. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1176.075942] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c0862-fd71-cd60-92bd-04815450d0c4/disk-0.vmdk. {{(pid=63418) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1176.076455] env[63418]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-690aff63-b2f7-455c-a18d-aca725852c19 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.250809] env[63418]: DEBUG oslo_vmware.rw_handles [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c0862-fd71-cd60-92bd-04815450d0c4/disk-0.vmdk. {{(pid=63418) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1176.251087] env[63418]: INFO nova.virt.vmwareapi.images [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Downloaded image file data cebca985-58be-495c-8736-51fa6a23dfda [ 1176.251973] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8311d292-9995-4fd1-87cb-02e6c734e47b {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.269309] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d22842b-2016-4055-ace2-1362d271dbdc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.305728] env[63418]: INFO nova.virt.vmwareapi.images [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] The imported VM was unregistered [ 1176.308144] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Caching image {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1176.308381] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Creating directory with path [datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.308644] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3920674d-5f95-4f65-bd93-3d8c5b914e9d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.319021] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Created directory with path [datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.319184] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260/OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260.vmdk to [datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk. {{(pid=63418) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1176.319414] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9d6405c0-9748-4f02-85b2-7cd39501e418 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.324963] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1176.324963] env[63418]: value = "task-1245657" [ 1176.324963] env[63418]: _type = "Task" [ 1176.324963] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.332707] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245657, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.835359] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245657, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.109833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.109833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.109833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.109833] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.110314] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.112454] env[63418]: INFO nova.compute.manager [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Terminating instance [ 1177.335622] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245657, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.616533] env[63418]: DEBUG nova.compute.manager [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1177.616689] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1177.617671] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0724e665-0bcb-4d04-8d62-f868d065b5f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.625625] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1177.626204] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ab061b7-84c4-4f44-ac9c-f4c14e0a7da8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.634322] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1177.634322] env[63418]: value = "task-1245658" [ 1177.634322] env[63418]: _type = "Task" [ 1177.634322] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.643528] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.836761] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245657, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.149584] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.337099] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245657, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.645196] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245658, 'name': PowerOffVM_Task, 'duration_secs': 0.936839} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.645494] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.645665] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1178.645928] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4af9d23e-c005-4f12-a6bc-771818a2cf7d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.725579] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1178.725841] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1178.726093] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleting the datastore file [datastore2] 38a131e4-22aa-41f1-8e7a-e2ab4660340c {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1178.726379] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e562bc0-a523-4d1b-ba9e-119407237f76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.732803] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1178.732803] env[63418]: value = "task-1245660" [ 1178.732803] env[63418]: _type = "Task" [ 1178.732803] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.740912] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245660, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.837815] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245657, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.277449} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.837815] env[63418]: INFO nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260/OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260.vmdk to [datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk. [ 1178.837815] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Cleaning up location [datastore2] OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1178.837958] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9477f3b6-e6d1-4097-af85-5ecea7843260 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1178.838253] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7993f448-9e9e-4ca6-86f5-aeffa2bdc0a8 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.844483] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1178.844483] env[63418]: value = "task-1245661" [ 1178.844483] env[63418]: _type = "Task" [ 1178.844483] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.852020] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245661, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.242153] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245660, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.355335] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245661, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332194} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.355655] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.355842] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.356124] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk to [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1179.356375] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca49d9d0-4a1c-4b33-9d7c-b09f56778a59 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.362044] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1179.362044] env[63418]: value = "task-1245662" [ 1179.362044] env[63418]: _type = "Task" [ 1179.362044] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.369265] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.743319] env[63418]: DEBUG oslo_vmware.api [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245660, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.621334} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.743646] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.743892] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1179.744101] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1179.744292] env[63418]: INFO nova.compute.manager [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Took 2.13 seconds to destroy the instance on the hypervisor. [ 1179.744544] env[63418]: DEBUG oslo.service.loopingcall [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1179.744756] env[63418]: DEBUG nova.compute.manager [-] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1179.744863] env[63418]: DEBUG nova.network.neutron [-] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1179.873079] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245662, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.306085] env[63418]: DEBUG nova.compute.manager [req-cbf3ec56-5e77-491d-8269-703b837fa94c req-34b7796e-a183-413f-bdc5-b4da89bfc473 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Received event network-vif-deleted-02a2964c-4337-419a-abc7-a73ef9a4af5f {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1180.306335] env[63418]: INFO nova.compute.manager [req-cbf3ec56-5e77-491d-8269-703b837fa94c req-34b7796e-a183-413f-bdc5-b4da89bfc473 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Neutron deleted interface 02a2964c-4337-419a-abc7-a73ef9a4af5f; detaching it from the instance and deleting it from the info cache [ 1180.306472] env[63418]: DEBUG nova.network.neutron [req-cbf3ec56-5e77-491d-8269-703b837fa94c req-34b7796e-a183-413f-bdc5-b4da89bfc473 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.374327] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245662, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.781913] env[63418]: DEBUG nova.network.neutron [-] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.810147] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2ac4a8a-2d1e-401d-b83a-8b12390af019 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.819954] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2502c739-ba52-4c64-8331-601af4144ebb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.843577] env[63418]: DEBUG nova.compute.manager [req-cbf3ec56-5e77-491d-8269-703b837fa94c req-34b7796e-a183-413f-bdc5-b4da89bfc473 service nova] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Detach interface failed, port_id=02a2964c-4337-419a-abc7-a73ef9a4af5f, reason: Instance 38a131e4-22aa-41f1-8e7a-e2ab4660340c could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1180.872834] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245662, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.210110] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.285250] env[63418]: INFO nova.compute.manager [-] [instance: 38a131e4-22aa-41f1-8e7a-e2ab4660340c] Took 1.54 seconds to deallocate network for instance. [ 1181.374403] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245662, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.792733] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.793034] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.793319] env[63418]: DEBUG nova.objects.instance [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'resources' on Instance uuid 38a131e4-22aa-41f1-8e7a-e2ab4660340c {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.875228] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245662, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.14974} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.875502] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cebca985-58be-495c-8736-51fa6a23dfda/cebca985-58be-495c-8736-51fa6a23dfda.vmdk to [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1181.876283] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6400c41-f25a-428a-b466-bd60dc19ebdf {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.899277] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.899504] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c289b5f-eac3-4ca3-8373-f7a23846809d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.917951] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1181.917951] env[63418]: value = "task-1245663" [ 1181.917951] env[63418]: _type = "Task" [ 1181.917951] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.925108] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245663, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.204843] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1182.346111] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83019c3-6a94-4caa-abf5-0c7bbd129e43 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.353432] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea332574-9940-46e2-bccf-6593e904ab2c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.381757] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e62b363-3493-4c13-a418-1b764ed309dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.389468] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20eb7b0f-2d7f-4645-8414-7b8066fccd00 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.401821] env[63418]: DEBUG nova.compute.provider_tree [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1182.425660] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245663, 'name': ReconfigVM_Task, 'duration_secs': 0.258555} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.426176] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70/95240ada-686c-484b-9c68-bc9a1f3d3e70.vmdk or device None with type streamOptimized {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.426778] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-54f6e366-2a88-4cef-9525-3a44468abfb5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.431983] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1182.431983] env[63418]: value = "task-1245664" [ 1182.431983] env[63418]: _type = "Task" [ 1182.431983] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.440019] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245664, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.905382] env[63418]: DEBUG nova.scheduler.client.report [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1182.942238] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245664, 'name': Rename_Task, 'duration_secs': 0.13072} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.942497] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.942736] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d456ff5-e6be-4d10-afd4-f8976027626c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.948957] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1182.948957] env[63418]: value = "task-1245665" [ 1182.948957] env[63418]: _type = "Task" [ 1182.948957] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.955969] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245665, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.410955] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.430853] env[63418]: INFO nova.scheduler.client.report [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted allocations for instance 38a131e4-22aa-41f1-8e7a-e2ab4660340c [ 1183.458072] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245665, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.938948] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2921f046-d44b-463f-9500-a517784cd9c9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "38a131e4-22aa-41f1-8e7a-e2ab4660340c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.829s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.958290] env[63418]: DEBUG oslo_vmware.api [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245665, 'name': PowerOnVM_Task, 'duration_secs': 0.522014} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.958551] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1184.054007] env[63418]: DEBUG nova.compute.manager [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1184.054927] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee63ac9-2f3f-4544-bc8b-41b96f39d8dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.209591] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.209759] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.572214] env[63418]: DEBUG oslo_concurrency.lockutils [None req-d03449a9-a144-4798-b398-bbf6ca263c81 tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.828s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.712790] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.713064] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.713277] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.713437] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1184.714384] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1244d3-18d6-43a2-85c4-ae82bf387ff5 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.722703] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980a4bf7-86d2-4d30-8b46-a08c857a5582 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.736250] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4acc571-bdd5-41da-baf5-fac3735428ba {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.742586] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12915bd-3984-496c-93fd-418ede9da238 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.773239] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181141MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1184.773409] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.773591] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.062596] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.062949] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.063218] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.063408] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.063585] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.066215] env[63418]: INFO nova.compute.manager [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Terminating instance [ 1185.570018] env[63418]: DEBUG nova.compute.manager [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1185.570018] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1185.570911] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9281c0e-7b6c-4c92-8fe6-d56d502e7283 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.579672] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.580551] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-593f1815-fcce-46d9-8ec3-3b7e022823eb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.587240] env[63418]: DEBUG oslo_vmware.api [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1185.587240] env[63418]: value = "task-1245666" [ 1185.587240] env[63418]: _type = "Task" [ 1185.587240] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.595597] env[63418]: DEBUG oslo_vmware.api [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245666, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.801613] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 95240ada-686c-484b-9c68-bc9a1f3d3e70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.802082] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1185.802392] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1185.831998] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac647336-4fc2-4bd1-98b6-73c2cb166239 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.839716] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494e60ff-e791-4985-8729-771112935ebe {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.871042] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b36f949-8fe6-45f6-8a31-5593bc1165da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.878397] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2800a095-ebaf-4627-a849-5deb9e1adada {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.891832] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.097478] env[63418]: DEBUG oslo_vmware.api [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245666, 'name': PowerOffVM_Task, 'duration_secs': 0.210111} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.097752] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.097923] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1186.098230] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e54ee1d-e2fa-4db3-a527-6b5a1c794fb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.164024] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1186.164303] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1186.164473] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleting the datastore file [datastore2] 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1186.164739] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55f68dce-3e03-4dd5-954c-dc7e771361dd {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.170677] env[63418]: DEBUG oslo_vmware.api [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for the task: (returnval){ [ 1186.170677] env[63418]: value = "task-1245668" [ 1186.170677] env[63418]: _type = "Task" [ 1186.170677] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.178079] env[63418]: DEBUG oslo_vmware.api [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245668, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.394458] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1186.401684] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.401911] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.681915] env[63418]: DEBUG oslo_vmware.api [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Task: {'id': task-1245668, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196744} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.682312] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.682439] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.682671] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.682887] env[63418]: INFO nova.compute.manager [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1186.683147] env[63418]: DEBUG oslo.service.loopingcall [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1186.683365] env[63418]: DEBUG nova.compute.manager [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1186.683464] env[63418]: DEBUG nova.network.neutron [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1186.904652] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1186.904652] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.129s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.904652] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Starting instance... {{(pid=63418) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1187.170614] env[63418]: DEBUG nova.compute.manager [req-298f46ec-97d7-4300-bc73-18136e9da213 req-f805aa44-7da5-45c4-8205-b1ba1d8d0ae8 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Received event network-vif-deleted-9d927b3d-91b4-429d-b74d-96a0e1822173 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1187.170831] env[63418]: INFO nova.compute.manager [req-298f46ec-97d7-4300-bc73-18136e9da213 req-f805aa44-7da5-45c4-8205-b1ba1d8d0ae8 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Neutron deleted interface 9d927b3d-91b4-429d-b74d-96a0e1822173; detaching it from the instance and deleting it from the info cache [ 1187.170831] env[63418]: DEBUG nova.network.neutron [req-298f46ec-97d7-4300-bc73-18136e9da213 req-f805aa44-7da5-45c4-8205-b1ba1d8d0ae8 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.425607] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.425878] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.427475] env[63418]: INFO nova.compute.claims [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1187.651206] env[63418]: DEBUG nova.network.neutron [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.673864] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9659a4ca-6182-4daf-8fce-63221ae88abb {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.683738] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47cf3ba-963c-458b-9d7a-5c209cca2548 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.706087] env[63418]: DEBUG nova.compute.manager [req-298f46ec-97d7-4300-bc73-18136e9da213 req-f805aa44-7da5-45c4-8205-b1ba1d8d0ae8 service nova] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Detach interface failed, port_id=9d927b3d-91b4-429d-b74d-96a0e1822173, reason: Instance 95240ada-686c-484b-9c68-bc9a1f3d3e70 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1187.902991] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.903241] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 1187.903370] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Rebuilding the list of instances to heal {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10282}} [ 1188.156052] env[63418]: INFO nova.compute.manager [-] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Took 1.47 seconds to deallocate network for instance. [ 1188.407081] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 95240ada-686c-484b-9c68-bc9a1f3d3e70] Skipping network cache update for instance because it is being deleted. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10295}} [ 1188.407081] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Skipping network cache update for instance because it is Building. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10291}} [ 1188.407081] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Didn't find any instances for network info cache update. {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10364}} [ 1188.407345] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.407345] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.407436] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.407556] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 1188.470952] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3289e5ce-7d82-4f51-a6e5-20f8fa4dbbf2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.478502] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca549bb-b269-4b4d-81a9-fcdb4830628e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.507787] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e92912a-f2e8-4e2e-8b08-27dc247ac714 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.514579] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa55f155-a6d0-471d-9117-23460b49ba76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.527009] env[63418]: DEBUG nova.compute.provider_tree [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.661183] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.709323] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.030062] env[63418]: DEBUG nova.scheduler.client.report [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1189.214369] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.534818] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.109s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.535360] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Start building networks asynchronously for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1189.538126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.877s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.538378] env[63418]: DEBUG nova.objects.instance [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lazy-loading 'resources' on Instance uuid 95240ada-686c-484b-9c68-bc9a1f3d3e70 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1190.041295] env[63418]: DEBUG nova.compute.utils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1190.045639] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Allocating IP information in the background. {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1190.045809] env[63418]: DEBUG nova.network.neutron [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] allocate_for_instance() {{(pid=63418) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1190.081623] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6ad149-936a-40a1-a5ca-36cabff65ca7 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.090014] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cfc8268-1095-4e4e-b0dd-8a2ef50db21f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.121280] env[63418]: DEBUG nova.policy [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b29941866349482fb9e53dcf87cb1845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05ef1c6c74574217817c6ab14a022b91', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63418) authorize /opt/stack/nova/nova/policy.py:201}} [ 1190.123169] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c3d190-b319-4107-bc88-4328f863d9f6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.130980] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca764a2-d559-4a68-94a4-96fcbc01b503 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.144044] env[63418]: DEBUG nova.compute.provider_tree [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.411947] env[63418]: DEBUG nova.network.neutron [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Successfully created port: 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1190.546827] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Start building block device mappings for instance. {{(pid=63418) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1190.648046] env[63418]: DEBUG nova.scheduler.client.report [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1191.152613] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.177978] env[63418]: INFO nova.scheduler.client.report [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Deleted allocations for instance 95240ada-686c-484b-9c68-bc9a1f3d3e70 [ 1191.557214] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Start spawning the instance on the hypervisor. {{(pid=63418) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1191.582436] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:36:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:35:45Z,direct_url=,disk_format='vmdk',id=c0dccf29-5e49-4a1e-b51f-d46e566b4772,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='41e71154fe804b74a68ae95330ed916e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.582690] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.582855] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image limits 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.583053] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Flavor pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.583212] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Image pref 0:0:0 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.583362] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63418) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.583575] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.583741] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.583913] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Got 1 possible topologies {{(pid=63418) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.584094] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.584310] env[63418]: DEBUG nova.virt.hardware [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63418) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.585152] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14907d2b-6a06-4e74-8834-55abd05c444c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.593013] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b534e175-65f9-4911-95f5-b432c5afecc2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.684897] env[63418]: DEBUG oslo_concurrency.lockutils [None req-177f28ce-0ff5-4e51-9094-233af4e93dbd tempest-ServerActionsTestOtherB-1129700828 tempest-ServerActionsTestOtherB-1129700828-project-member] Lock "95240ada-686c-484b-9c68-bc9a1f3d3e70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.622s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.796748] env[63418]: DEBUG nova.compute.manager [req-1dcc391a-9841-4fd1-9c32-e3b9d88c1489 req-5855780e-c963-4270-8d6b-271ddaa3f654 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Received event network-vif-plugged-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1191.796975] env[63418]: DEBUG oslo_concurrency.lockutils [req-1dcc391a-9841-4fd1-9c32-e3b9d88c1489 req-5855780e-c963-4270-8d6b-271ddaa3f654 service nova] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.797264] env[63418]: DEBUG oslo_concurrency.lockutils [req-1dcc391a-9841-4fd1-9c32-e3b9d88c1489 req-5855780e-c963-4270-8d6b-271ddaa3f654 service nova] Lock "73249272-239b-453d-8c14-1e78a233b3f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.797374] env[63418]: DEBUG oslo_concurrency.lockutils [req-1dcc391a-9841-4fd1-9c32-e3b9d88c1489 req-5855780e-c963-4270-8d6b-271ddaa3f654 service nova] Lock "73249272-239b-453d-8c14-1e78a233b3f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.797547] env[63418]: DEBUG nova.compute.manager [req-1dcc391a-9841-4fd1-9c32-e3b9d88c1489 req-5855780e-c963-4270-8d6b-271ddaa3f654 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] No waiting events found dispatching network-vif-plugged-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1191.797716] env[63418]: WARNING nova.compute.manager [req-1dcc391a-9841-4fd1-9c32-e3b9d88c1489 req-5855780e-c963-4270-8d6b-271ddaa3f654 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Received unexpected event network-vif-plugged-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 for instance with vm_state building and task_state spawning. [ 1191.885190] env[63418]: DEBUG nova.network.neutron [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Successfully updated port: 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1192.387959] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.388295] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.388295] env[63418]: DEBUG nova.network.neutron [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Building network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1192.918801] env[63418]: DEBUG nova.network.neutron [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Instance cache missing network info. {{(pid=63418) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1193.043019] env[63418]: DEBUG nova.network.neutron [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating instance_info_cache with network_info: [{"id": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "address": "fa:16:3e:b6:ed:6f", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429703ac-ba", "ovs_interfaceid": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.546013] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.546371] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Instance network_info: |[{"id": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "address": "fa:16:3e:b6:ed:6f", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429703ac-ba", "ovs_interfaceid": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63418) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1193.546818] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:ed:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'af454577-0e89-41a3-a9f2-f39716f62fd5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6', 'vif_model': 'vmxnet3'}] {{(pid=63418) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1193.554731] env[63418]: DEBUG oslo.service.loopingcall [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1193.554944] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Creating VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1193.555185] env[63418]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-499c9f39-4090-4369-a1c7-c1e063b97181 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.576525] env[63418]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1193.576525] env[63418]: value = "task-1245670" [ 1193.576525] env[63418]: _type = "Task" [ 1193.576525] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.583963] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245670, 'name': CreateVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.822269] env[63418]: DEBUG nova.compute.manager [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Received event network-changed-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1193.822476] env[63418]: DEBUG nova.compute.manager [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Refreshing instance network info cache due to event network-changed-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1193.822691] env[63418]: DEBUG oslo_concurrency.lockutils [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] Acquiring lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.822838] env[63418]: DEBUG oslo_concurrency.lockutils [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] Acquired lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.823011] env[63418]: DEBUG nova.network.neutron [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Refreshing network info cache for port 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1194.086624] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245670, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.557416] env[63418]: DEBUG nova.network.neutron [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updated VIF entry in instance network info cache for port 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1194.557826] env[63418]: DEBUG nova.network.neutron [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating instance_info_cache with network_info: [{"id": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "address": "fa:16:3e:b6:ed:6f", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429703ac-ba", "ovs_interfaceid": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.587193] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245670, 'name': CreateVM_Task} progress is 99%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.061175] env[63418]: DEBUG oslo_concurrency.lockutils [req-699fb167-0c57-4b95-95f2-0cd59c34c67e req-9303c59e-bd55-456f-ab64-f0753e08e543 service nova] Releasing lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.088174] env[63418]: DEBUG oslo_vmware.api [-] Task: {'id': task-1245670, 'name': CreateVM_Task, 'duration_secs': 1.303868} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.088390] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Created VM on the ESX host {{(pid=63418) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1195.089103] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.089304] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.089628] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1195.089923] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36171385-6379-48e0-acb6-4ce7aa6cf270 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.094479] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1195.094479] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525884e3-e4ce-d71b-d4d3-db427a1c993f" [ 1195.094479] env[63418]: _type = "Task" [ 1195.094479] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.101595] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525884e3-e4ce-d71b-d4d3-db427a1c993f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.605310] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525884e3-e4ce-d71b-d4d3-db427a1c993f, 'name': SearchDatastore_Task, 'duration_secs': 0.010689} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.605682] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.605871] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Processing image c0dccf29-5e49-4a1e-b51f-d46e566b4772 {{(pid=63418) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1195.606159] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.606285] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.606491] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1195.606831] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0bc480c-7870-45e6-820b-8d07c2d3078d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.614968] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63418) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1195.615172] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63418) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1195.615896] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6be6c658-ec78-4385-8e08-ab20607d5588 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.621280] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1195.621280] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]525cbc41-ad86-825f-2c4d-6f5f7e4deba3" [ 1195.621280] env[63418]: _type = "Task" [ 1195.621280] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.630030] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525cbc41-ad86-825f-2c4d-6f5f7e4deba3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.131938] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]525cbc41-ad86-825f-2c4d-6f5f7e4deba3, 'name': SearchDatastore_Task, 'duration_secs': 0.008484} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.132838] env[63418]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2be079c5-be70-440a-8060-c46e02e7648d {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.139213] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1196.139213] env[63418]: value = "session[5246eacd-9b33-25a4-4179-4bd40204edb0]5263c7ec-7921-46e4-61d1-acb1620246b9" [ 1196.139213] env[63418]: _type = "Task" [ 1196.139213] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.147213] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5263c7ec-7921-46e4-61d1-acb1620246b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.651036] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': session[5246eacd-9b33-25a4-4179-4bd40204edb0]5263c7ec-7921-46e4-61d1-acb1620246b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009176} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.651036] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.651036] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 73249272-239b-453d-8c14-1e78a233b3f3/73249272-239b-453d-8c14-1e78a233b3f3.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1196.651036] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61865bdb-cba8-4247-ac6a-c9c46bd28264 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.656682] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1196.656682] env[63418]: value = "task-1245671" [ 1196.656682] env[63418]: _type = "Task" [ 1196.656682] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.663840] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245671, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.167514] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245671, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4449} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.167815] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0dccf29-5e49-4a1e-b51f-d46e566b4772/c0dccf29-5e49-4a1e-b51f-d46e566b4772.vmdk to [datastore2] 73249272-239b-453d-8c14-1e78a233b3f3/73249272-239b-453d-8c14-1e78a233b3f3.vmdk {{(pid=63418) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1197.168047] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Extending root virtual disk to 1048576 {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1197.168332] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b31050e9-18ea-4a00-b95d-34482f2414d0 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.174886] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1197.174886] env[63418]: value = "task-1245672" [ 1197.174886] env[63418]: _type = "Task" [ 1197.174886] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.181941] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.684238] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074879} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.684620] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Extended root virtual disk {{(pid=63418) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1197.685260] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170d72ec-68cd-4a76-86e7-b91eda3b6279 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.706537] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 73249272-239b-453d-8c14-1e78a233b3f3/73249272-239b-453d-8c14-1e78a233b3f3.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1197.706773] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1850e141-4d38-49fc-b049-ebc789cd3737 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.726435] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1197.726435] env[63418]: value = "task-1245673" [ 1197.726435] env[63418]: _type = "Task" [ 1197.726435] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.733456] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.236785] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245673, 'name': ReconfigVM_Task, 'duration_secs': 0.341987} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.237033] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 73249272-239b-453d-8c14-1e78a233b3f3/73249272-239b-453d-8c14-1e78a233b3f3.vmdk or device None with type sparse {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1198.237644] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f89e4e04-ab09-4a41-881c-4ce640915fd2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.243099] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1198.243099] env[63418]: value = "task-1245674" [ 1198.243099] env[63418]: _type = "Task" [ 1198.243099] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.250604] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245674, 'name': Rename_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.752699] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245674, 'name': Rename_Task, 'duration_secs': 0.140558} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.753071] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Powering on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1198.753229] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f1cf434-c407-47eb-9528-d6069dfdf640 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.759284] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1198.759284] env[63418]: value = "task-1245675" [ 1198.759284] env[63418]: _type = "Task" [ 1198.759284] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.767551] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.270652] env[63418]: DEBUG oslo_vmware.api [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245675, 'name': PowerOnVM_Task, 'duration_secs': 0.44539} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.270924] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Powered on the VM {{(pid=63418) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1199.271147] env[63418]: INFO nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1199.271331] env[63418]: DEBUG nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Checking state {{(pid=63418) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1199.272154] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e70810-85d5-4547-9f4e-d7056e73dd0c {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.789524] env[63418]: INFO nova.compute.manager [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Took 12.38 seconds to build instance. [ 1200.291718] env[63418]: DEBUG oslo_concurrency.lockutils [None req-6b0c5ce4-cdb7-4b76-8a86-7d4ac9b25ca9 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.890s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.690724] env[63418]: DEBUG nova.compute.manager [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Received event network-changed-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1200.690949] env[63418]: DEBUG nova.compute.manager [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Refreshing instance network info cache due to event network-changed-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6. {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11486}} [ 1200.691182] env[63418]: DEBUG oslo_concurrency.lockutils [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] Acquiring lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.691329] env[63418]: DEBUG oslo_concurrency.lockutils [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] Acquired lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.691492] env[63418]: DEBUG nova.network.neutron [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Refreshing network info cache for port 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1201.392132] env[63418]: DEBUG nova.network.neutron [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updated VIF entry in instance network info cache for port 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6. {{(pid=63418) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1201.392538] env[63418]: DEBUG nova.network.neutron [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating instance_info_cache with network_info: [{"id": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "address": "fa:16:3e:b6:ed:6f", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429703ac-ba", "ovs_interfaceid": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.895476] env[63418]: DEBUG oslo_concurrency.lockutils [req-c0b304c1-42ab-4401-b7b3-4d88441c5028 req-9ded713b-a722-485d-ac1b-926985dc3112 service nova] Releasing lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.041067] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.041353] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.544447] env[63418]: DEBUG nova.compute.utils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Using /dev/sd instead of None {{(pid=63418) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1240.047514] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.143894] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.144344] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.144480] env[63418]: INFO nova.compute.manager [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Attaching volume 54987088-edb0-49ed-be04-0a70d67aff33 to /dev/sdb [ 1241.174059] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ce38fa-6c67-44bc-b38a-f2d7df62fc05 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.181383] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6595409c-727f-4354-a592-4c5b15e876ff {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.194076] env[63418]: DEBUG nova.virt.block_device [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating existing volume attachment record: a32627c8-84d5-402d-b918-42fc0320f35c {{(pid=63418) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1242.709968] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1243.209778] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1245.736638] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Volume attach. Driver type: vmdk {{(pid=63418) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1245.736902] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268560', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'name': 'volume-54987088-edb0-49ed-be04-0a70d67aff33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '73249272-239b-453d-8c14-1e78a233b3f3', 'attached_at': '', 'detached_at': '', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'serial': '54987088-edb0-49ed-be04-0a70d67aff33'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1245.737803] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9c8dc5-d9ce-40ba-8ad7-270d5aeb347a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.753660] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1124ec16-fa70-4d64-8a67-b419f20aab8f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.776585] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] volume-54987088-edb0-49ed-be04-0a70d67aff33/volume-54987088-edb0-49ed-be04-0a70d67aff33.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1245.776822] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac58481f-6247-44e5-90c6-8f793b4da7bc {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.793972] env[63418]: DEBUG oslo_vmware.api [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1245.793972] env[63418]: value = "task-1245678" [ 1245.793972] env[63418]: _type = "Task" [ 1245.793972] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.801009] env[63418]: DEBUG oslo_vmware.api [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245678, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.210209] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.210447] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Starting heal instance info cache {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10278}} [ 1246.210513] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Rebuilding the list of instances to heal {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10282}} [ 1246.303930] env[63418]: DEBUG oslo_vmware.api [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245678, 'name': ReconfigVM_Task, 'duration_secs': 0.305307} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.304251] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Reconfigured VM instance instance-00000071 to attach disk [datastore2] volume-54987088-edb0-49ed-be04-0a70d67aff33/volume-54987088-edb0-49ed-be04-0a70d67aff33.vmdk or device None with type thin {{(pid=63418) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1246.308920] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25963e92-80fc-44b5-a272-818a4ad9b342 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.323340] env[63418]: DEBUG oslo_vmware.api [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1246.323340] env[63418]: value = "task-1245679" [ 1246.323340] env[63418]: _type = "Task" [ 1246.323340] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.330664] env[63418]: DEBUG oslo_vmware.api [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245679, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.752942] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1246.753231] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquired lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1246.753498] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Forcefully refreshing network info cache for instance {{(pid=63418) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1246.753772] env[63418]: DEBUG nova.objects.instance [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lazy-loading 'info_cache' on Instance uuid 73249272-239b-453d-8c14-1e78a233b3f3 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1246.832940] env[63418]: DEBUG oslo_vmware.api [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245679, 'name': ReconfigVM_Task, 'duration_secs': 0.127226} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.833288] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268560', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'name': 'volume-54987088-edb0-49ed-be04-0a70d67aff33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '73249272-239b-453d-8c14-1e78a233b3f3', 'attached_at': '', 'detached_at': '', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'serial': '54987088-edb0-49ed-be04-0a70d67aff33'} {{(pid=63418) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1247.870035] env[63418]: DEBUG nova.objects.instance [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'flavor' on Instance uuid 73249272-239b-453d-8c14-1e78a233b3f3 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1248.379687] env[63418]: DEBUG oslo_concurrency.lockutils [None req-4f7b1d2c-8d91-4832-96ad-4dc76ff462ab tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.482705] env[63418]: DEBUG nova.network.neutron [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating instance_info_cache with network_info: [{"id": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "address": "fa:16:3e:b6:ed:6f", "network": {"id": "030a6ba8-285a-4639-be43-14047c545ec8", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1893506199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05ef1c6c74574217817c6ab14a022b91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "af454577-0e89-41a3-a9f2-f39716f62fd5", "external-id": "nsx-vlan-transportzone-63", "segmentation_id": 63, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429703ac-ba", "ovs_interfaceid": "429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.583906] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.584183] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.985460] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Releasing lock "refresh_cache-73249272-239b-453d-8c14-1e78a233b3f3" {{(pid=63418) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1248.985891] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updated the network info_cache for instance {{(pid=63418) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10349}} [ 1248.985891] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.986022] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.986181] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.986311] env[63418]: DEBUG nova.compute.manager [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63418) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10897}} [ 1248.986452] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.087436] env[63418]: INFO nova.compute.manager [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Detaching volume 54987088-edb0-49ed-be04-0a70d67aff33 [ 1249.116862] env[63418]: INFO nova.virt.block_device [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Attempting to driver detach volume 54987088-edb0-49ed-be04-0a70d67aff33 from mountpoint /dev/sdb [ 1249.117156] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Volume detach. Driver type: vmdk {{(pid=63418) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1249.117361] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268560', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'name': 'volume-54987088-edb0-49ed-be04-0a70d67aff33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '73249272-239b-453d-8c14-1e78a233b3f3', 'attached_at': '', 'detached_at': '', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'serial': '54987088-edb0-49ed-be04-0a70d67aff33'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1249.118230] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27fa017-f109-49dc-8463-12875c340da3 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.139554] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50940320-e741-4a4c-aace-f7d6ecdb289f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.145970] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913b9244-17f3-4fd6-9bf3-8be803bf20da {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.166807] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1bb14c-8808-49b9-b380-efed6532d530 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.180454] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] The volume has not been displaced from its original location: [datastore2] volume-54987088-edb0-49ed-be04-0a70d67aff33/volume-54987088-edb0-49ed-be04-0a70d67aff33.vmdk. No consolidation needed. {{(pid=63418) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1249.185675] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Reconfiguring VM instance instance-00000071 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1249.185927] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-416a26e7-5df3-46fa-a3b1-2096167a5328 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.202183] env[63418]: DEBUG oslo_vmware.api [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1249.202183] env[63418]: value = "task-1245680" [ 1249.202183] env[63418]: _type = "Task" [ 1249.202183] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.209036] env[63418]: DEBUG oslo_vmware.api [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245680, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.490165] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.490447] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.490581] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.490730] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63418) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1249.491653] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053d4228-0d53-46e1-9d28-f07800ae633a {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.499345] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b429a842-0833-4a2a-beb5-440ea4505477 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.512594] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7bafb0-15c2-41a9-9200-e405f19b7ab2 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.518714] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f892dac-601c-4b52-83a2-8ec64f776089 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.547445] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181298MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63418) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1249.547580] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.547765] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.711085] env[63418]: DEBUG oslo_vmware.api [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245680, 'name': ReconfigVM_Task, 'duration_secs': 0.202763} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.711377] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Reconfigured VM instance instance-00000071 to detach disk 2001 {{(pid=63418) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1249.715945] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6767b0d0-2196-4a77-9b0c-71366826390e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.730514] env[63418]: DEBUG oslo_vmware.api [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1249.730514] env[63418]: value = "task-1245681" [ 1249.730514] env[63418]: _type = "Task" [ 1249.730514] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.739014] env[63418]: DEBUG oslo_vmware.api [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.240140] env[63418]: DEBUG oslo_vmware.api [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245681, 'name': ReconfigVM_Task, 'duration_secs': 0.136426} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.240484] env[63418]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268560', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'name': 'volume-54987088-edb0-49ed-be04-0a70d67aff33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '73249272-239b-453d-8c14-1e78a233b3f3', 'attached_at': '', 'detached_at': '', 'volume_id': '54987088-edb0-49ed-be04-0a70d67aff33', 'serial': '54987088-edb0-49ed-be04-0a70d67aff33'} {{(pid=63418) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1250.573411] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Instance 73249272-239b-453d-8c14-1e78a233b3f3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63418) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.573612] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1250.573756] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63418) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1250.599960] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4611e203-4daa-418a-b7b2-ffbd495be6ec {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.607596] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90703ca3-3939-488b-bbf4-403554ffba4f {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.636290] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6a23fc-ba67-4b48-bf46-f2c5b3b23b76 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.643208] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea0021c-eda1-421d-b595-51c6e63a77a6 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.655684] env[63418]: DEBUG nova.compute.provider_tree [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1250.779916] env[63418]: DEBUG nova.objects.instance [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'flavor' on Instance uuid 73249272-239b-453d-8c14-1e78a233b3f3 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1251.159010] env[63418]: DEBUG nova.scheduler.client.report [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1251.663717] env[63418]: DEBUG nova.compute.resource_tracker [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63418) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1251.664126] env[63418]: DEBUG oslo_concurrency.lockutils [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.787029] env[63418]: DEBUG oslo_concurrency.lockutils [None req-2d700506-70bb-43d9-8411-0a896c870e0d tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.203s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.820161] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.820629] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.820808] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "73249272-239b-453d-8c14-1e78a233b3f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.821013] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.821196] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.823413] env[63418]: INFO nova.compute.manager [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Terminating instance [ 1252.887386] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1252.887628] env[63418]: DEBUG oslo_service.periodic_task [None req-7e78b118-077b-4e8f-8a32-2ac0b25560d1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63418) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.327430] env[63418]: DEBUG nova.compute.manager [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Start destroying the instance on the hypervisor. {{(pid=63418) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1253.327626] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Destroying instance {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1253.328549] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265d09c9-e881-4272-b43d-ae056cc0c6f4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.336354] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Powering off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1253.336584] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edac9d9a-470b-4cdd-b60c-53303d3a5cb1 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.342732] env[63418]: DEBUG oslo_vmware.api [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1253.342732] env[63418]: value = "task-1245682" [ 1253.342732] env[63418]: _type = "Task" [ 1253.342732] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.350618] env[63418]: DEBUG oslo_vmware.api [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.852598] env[63418]: DEBUG oslo_vmware.api [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245682, 'name': PowerOffVM_Task, 'duration_secs': 0.16576} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.852973] env[63418]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Powered off the VM {{(pid=63418) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1253.853079] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Unregistering the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1253.853278] env[63418]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85a53815-3381-4a17-8016-18644d10fece {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.916044] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Unregistered the VM {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1253.916296] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Deleting contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1253.916486] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleting the datastore file [datastore2] 73249272-239b-453d-8c14-1e78a233b3f3 {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1253.916752] env[63418]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f8e6781-4574-4c8d-a177-37d8abfc6c7e {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.923097] env[63418]: DEBUG oslo_vmware.api [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for the task: (returnval){ [ 1253.923097] env[63418]: value = "task-1245684" [ 1253.923097] env[63418]: _type = "Task" [ 1253.923097] env[63418]: } to complete. {{(pid=63418) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.929968] env[63418]: DEBUG oslo_vmware.api [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.433353] env[63418]: DEBUG oslo_vmware.api [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Task: {'id': task-1245684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128104} completed successfully. {{(pid=63418) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.433625] env[63418]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted the datastore file {{(pid=63418) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1254.433818] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Deleted contents of the VM from datastore datastore2 {{(pid=63418) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1254.433998] env[63418]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Instance destroyed {{(pid=63418) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1254.434197] env[63418]: INFO nova.compute.manager [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1254.434451] env[63418]: DEBUG oslo.service.loopingcall [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63418) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1254.434643] env[63418]: DEBUG nova.compute.manager [-] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Deallocating network for instance {{(pid=63418) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1254.434741] env[63418]: DEBUG nova.network.neutron [-] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] deallocate_for_instance() {{(pid=63418) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1254.883132] env[63418]: DEBUG nova.compute.manager [req-f48f5360-94a3-4174-a751-14997b0a96f7 req-04494552-b982-4125-9d28-d3834911ebee service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Received event network-vif-deleted-429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6 {{(pid=63418) external_instance_event /opt/stack/nova/nova/compute/manager.py:11481}} [ 1254.883132] env[63418]: INFO nova.compute.manager [req-f48f5360-94a3-4174-a751-14997b0a96f7 req-04494552-b982-4125-9d28-d3834911ebee service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Neutron deleted interface 429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6; detaching it from the instance and deleting it from the info cache [ 1254.883132] env[63418]: DEBUG nova.network.neutron [req-f48f5360-94a3-4174-a751-14997b0a96f7 req-04494552-b982-4125-9d28-d3834911ebee service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.363893] env[63418]: DEBUG nova.network.neutron [-] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Updating instance_info_cache with network_info: [] {{(pid=63418) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.385541] env[63418]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c44e4530-1f38-4133-a92d-13b5286b5fda {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.395127] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b110a5-f2c0-40e5-af65-82b5caa4aa16 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.418700] env[63418]: DEBUG nova.compute.manager [req-f48f5360-94a3-4174-a751-14997b0a96f7 req-04494552-b982-4125-9d28-d3834911ebee service nova] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Detach interface failed, port_id=429703ac-ba13-43f5-aaf2-ba5b8b3b5bc6, reason: Instance 73249272-239b-453d-8c14-1e78a233b3f3 could not be found. {{(pid=63418) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11315}} [ 1255.866339] env[63418]: INFO nova.compute.manager [-] [instance: 73249272-239b-453d-8c14-1e78a233b3f3] Took 1.43 seconds to deallocate network for instance. [ 1256.372877] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.373269] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.373435] env[63418]: DEBUG nova.objects.instance [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lazy-loading 'resources' on Instance uuid 73249272-239b-453d-8c14-1e78a233b3f3 {{(pid=63418) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.908052] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21efffe3-d4fa-4e09-a32a-2feb448f7343 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.915432] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcb7887-8f41-40d9-b693-cfb64b1db547 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.944932] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fc4308-6224-4d65-b320-ecd34d9d7981 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.952351] env[63418]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64119724-9a17-4d88-8a37-b941bfbc0eb4 {{(pid=63418) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.964983] env[63418]: DEBUG nova.compute.provider_tree [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed in ProviderTree for provider: 6ac9de28-4c58-4fc2-8a3d-711092e3c63c {{(pid=63418) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1257.467956] env[63418]: DEBUG nova.scheduler.client.report [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Inventory has not changed for provider 6ac9de28-4c58-4fc2-8a3d-711092e3c63c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63418) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1257.973111] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.994893] env[63418]: INFO nova.scheduler.client.report [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Deleted allocations for instance 73249272-239b-453d-8c14-1e78a233b3f3 [ 1258.504635] env[63418]: DEBUG oslo_concurrency.lockutils [None req-9d3b4fa4-5a97-4927-be28-0d84f3861988 tempest-AttachVolumeNegativeTest-64398399 tempest-AttachVolumeNegativeTest-64398399-project-member] Lock "73249272-239b-453d-8c14-1e78a233b3f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.684s {{(pid=63418) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}